Search is not available for this dataset
url string | repository_url string | labels_url string | comments_url string | events_url string | html_url string | id int64 | node_id string | number int64 | title string | user dict | labels list | state string | locked bool | assignee dict | assignees list | milestone dict | comments list | created_at string | updated_at string | closed_at string | author_association string | type null | active_lock_reason null | sub_issues_summary dict | body string | closed_by dict | reactions dict | timeline_url string | performed_via_github_app null | state_reason string | draft bool | pull_request dict |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
https://api.github.com/repos/huggingface/datasets/issues/7686 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7686/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7686/comments | https://api.github.com/repos/huggingface/datasets/issues/7686/events | https://github.com/huggingface/datasets/issues/7686 | 3,237,201,090 | I_kwDODunzps7A88TC | 7,686 | load_dataset does not check .no_exist files in the hub cache | {
"avatar_url": "https://avatars.githubusercontent.com/u/3627235?v=4",
"events_url": "https://api.github.com/users/jmaccarl/events{/privacy}",
"followers_url": "https://api.github.com/users/jmaccarl/followers",
"following_url": "https://api.github.com/users/jmaccarl/following{/other_user}",
"gists_url": "https://api.github.com/users/jmaccarl/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/jmaccarl",
"id": 3627235,
"login": "jmaccarl",
"node_id": "MDQ6VXNlcjM2MjcyMzU=",
"organizations_url": "https://api.github.com/users/jmaccarl/orgs",
"received_events_url": "https://api.github.com/users/jmaccarl/received_events",
"repos_url": "https://api.github.com/users/jmaccarl/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/jmaccarl/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmaccarl/subscriptions",
"type": "User",
"url": "https://api.github.com/users/jmaccarl",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [] | 2025-07-16T20:04:00Z | 2025-07-16T20:04:00Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
I'm not entirely sure if this should be submitted as a bug in the `datasets` library or the `huggingface_hub` library, given it could be fixed at different levels of the stack.
The fundamental issue is that the `load_datasets` api doesn't use the `.no_exist` files in the hub cache unlike other wrapper APIs that do. This is because the `utils.file_utils.cached_path` used directly calls `hf_hub_download` instead of using `file_download.try_to_load_from_cache` from `huggingface_hub` (see `transformers` library `utils.hub.cached_files` for one alternate example).
This results in unnecessary metadata HTTP requests occurring for files that don't exist on every call. It won't generate the .no_exist cache files, nor will it use them.
### Steps to reproduce the bug
Run the following snippet as one example (setting cache dirs to clean paths for clarity)
`env HF_HOME=~/local_hf_hub python repro.py`
```
from datasets import load_dataset
import huggingface_hub
# monkeypatch to print out metadata requests being made
original_get_hf_file_metadata = huggingface_hub.file_download.get_hf_file_metadata
def get_hf_file_metadata_wrapper(*args, **kwargs):
print("File metadata request made (get_hf_file_metadata):", args, kwargs)
return original_get_hf_file_metadata(*args, **kwargs)
# Apply the patch
huggingface_hub.file_download.get_hf_file_metadata = get_hf_file_metadata_wrapper
dataset = load_dataset(
"Salesforce/wikitext",
"wikitext-2-v1",
split="test",
trust_remote_code=True,
cache_dir="~/local_datasets",
revision="b08601e04326c79dfdd32d625aee71d232d685c3",
)
```
This may be called over and over again, and you will see the same calls for files that don't exist:
```
File metadata request made (get_hf_file_metadata): () {'url': 'https://huggingface.co/datasets/Salesforce/wikitext/resolve/b08601e04326c79dfdd32d625aee71d232d685c3/wikitext.py', 'proxies': None, 'timeout': 10, 'headers': {'user-agent': 'datasets/3.6.0; hf_hub/0.33.2; python/3.12.11; torch/2.7.0; huggingface_hub/0.33.2; pyarrow/20.0.0; jax/0.5.3'}, 'token': None}
File metadata request made (get_hf_file_metadata): () {'url': 'https://huggingface.co/datasets/Salesforce/wikitext/resolve/b08601e04326c79dfdd32d625aee71d232d685c3/.huggingface.yaml', 'proxies': None, 'timeout': 10, 'headers': {'user-agent': 'datasets/3.6.0; hf_hub/0.33.2; python/3.12.11; torch/2.7.0; huggingface_hub/0.33.2; pyarrow/20.0.0; jax/0.5.3'}, 'token': None}
File metadata request made (get_hf_file_metadata): () {'url': 'https://huggingface.co/datasets/Salesforce/wikitext/resolve/b08601e04326c79dfdd32d625aee71d232d685c3/dataset_infos.json', 'proxies': None, 'timeout': 10, 'headers': {'user-agent': 'datasets/3.6.0; hf_hub/0.33.2; python/3.12.11; torch/2.7.0; huggingface_hub/0.33.2; pyarrow/20.0.0; jax/0.5.3'}, 'token': None}
```
And you can see that the .no_exist folder is never created
```
$ ls ~/local_hf_hub/hub/datasets--Salesforce--wikitext/
blobs refs snapshots
```
### Expected behavior
The expected behavior is for the print "File metadata request made" to stop after the first call, and for .no_exist directory & files to be populated under ~/local_hf_hub/hub/datasets--Salesforce--wikitext/
### Environment info
- `datasets` version: 3.6.0
- Platform: Linux-6.5.13-65-650-4141-22041-coreweave-amd64-85c45edc-x86_64-with-glibc2.35
- Python version: 3.12.11
- `huggingface_hub` version: 0.33.2
- PyArrow version: 20.0.0
- Pandas version: 2.3.1
- `fsspec` version: 2024.9.0 | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7686/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7686/timeline | null | null | null | null |
https://api.github.com/repos/huggingface/datasets/issues/7685 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7685/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7685/comments | https://api.github.com/repos/huggingface/datasets/issues/7685/events | https://github.com/huggingface/datasets/issues/7685 | 3,236,979,340 | I_kwDODunzps7A8GKM | 7,685 | Inconsistent range request behavior for parquet REST api | {
"avatar_url": "https://avatars.githubusercontent.com/u/21327470?v=4",
"events_url": "https://api.github.com/users/universalmind303/events{/privacy}",
"followers_url": "https://api.github.com/users/universalmind303/followers",
"following_url": "https://api.github.com/users/universalmind303/following{/other_user}",
"gists_url": "https://api.github.com/users/universalmind303/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/universalmind303",
"id": 21327470,
"login": "universalmind303",
"node_id": "MDQ6VXNlcjIxMzI3NDcw",
"organizations_url": "https://api.github.com/users/universalmind303/orgs",
"received_events_url": "https://api.github.com/users/universalmind303/received_events",
"repos_url": "https://api.github.com/users/universalmind303/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/universalmind303/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/universalmind303/subscriptions",
"type": "User",
"url": "https://api.github.com/users/universalmind303",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [] | 2025-07-16T18:39:44Z | 2025-07-16T18:41:53Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
First off, I do apologize if this is not the correct repo for submitting this issue. Please direct me to another one if it's more appropriate elsewhere.
The datasets rest api is inconsistently giving `416 Range Not Satisfiable` when using a range request to get portions of the parquet files. More often than not, I am seeing 416, but other times for an identical request, it gives me the data along with `206 Partial Content` as expected.
### Steps to reproduce the bug
repeating this request multiple times will return either 416 or 206.
```sh
$ curl -v -L -H "Range: bytes=217875070-218006142" -o output.parquet "https://huggingface.co/api/datasets/HuggingFaceTB/smoltalk2/parquet/Mid/Llama_Nemotron_Post_Training_Dataset_reasoning_r1/0.parquet"
```
Note: this is not limited to just the above file, I tried with many different datasets and am able to consistently reproduce issue across multiple datasets.
when the 416 is returned, I get the following headers
```
< HTTP/2 416
< content-type: text/html
< content-length: 49
< server: CloudFront
< date: Wed, 16 Jul 2025 14:58:43 GMT
< expires: Wed, 16 Jul 2025 14:58:43 GMT
< content-range: bytes */177
< x-cache: Error from cloudfront
< via: 1.1 873527676a354c5998cad133525df9c0.cloudfront.net (CloudFront)
<
```
this suggests to me that there is likely a CDN/caching/routing issue happening and the request is not getting routed properly.
Full verbose output via curl.
<details>
❯ curl -v -L -H "Range: bytes=217875070-218006142" -o output.parquet "https://huggingface.co/api/datasets/HuggingFaceTB/smoltalk2/parquet/Mid/Llama_Nemotron_Post_Training_Dataset_reasoning_r1/0.parquet"
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0* Host huggingface.co:443 was resolved.
* IPv6: (none)
* IPv4: 18.160.102.96, 18.160.102.110, 18.160.102.4, 18.160.102.86
* Trying 18.160.102.96:443...
* Connected to huggingface.co (18.160.102.96) port 443
* ALPN: curl offers h2,http/1.1
* (304) (OUT), TLS handshake, Client hello (1):
} [319 bytes data]
* CAfile: /etc/ssl/cert.pem
* CApath: none
* (304) (IN), TLS handshake, Server hello (2):
{ [122 bytes data]
* (304) (IN), TLS handshake, Unknown (8):
{ [19 bytes data]
* (304) (IN), TLS handshake, Certificate (11):
{ [3821 bytes data]
* (304) (IN), TLS handshake, CERT verify (15):
{ [264 bytes data]
* (304) (IN), TLS handshake, Finished (20):
{ [36 bytes data]
* (304) (OUT), TLS handshake, Finished (20):
} [36 bytes data]
* SSL connection using TLSv1.3 / AEAD-AES128-GCM-SHA256 / [blank] / UNDEF
* ALPN: server accepted h2
* Server certificate:
* subject: CN=huggingface.co
* start date: Apr 13 00:00:00 2025 GMT
* expire date: May 12 23:59:59 2026 GMT
* subjectAltName: host "huggingface.co" matched cert's "huggingface.co"
* issuer: C=US; O=Amazon; CN=Amazon RSA 2048 M02
* SSL certificate verify ok.
* using HTTP/2
* [HTTP/2] [1] OPENED stream for https://huggingface.co/api/datasets/HuggingFaceTB/smoltalk2/parquet/Mid/Llama_Nemotron_Post_Training_Dataset_reasoning_r1/0.parquet
* [HTTP/2] [1] [:method: GET]
* [HTTP/2] [1] [:scheme: https]
* [HTTP/2] [1] [:authority: huggingface.co]
* [HTTP/2] [1] [:path: /api/datasets/HuggingFaceTB/smoltalk2/parquet/Mid/Llama_Nemotron_Post_Training_Dataset_reasoning_r1/0.parquet]
* [HTTP/2] [1] [user-agent: curl/8.7.1]
* [HTTP/2] [1] [accept: */*]
* [HTTP/2] [1] [range: bytes=217875070-218006142]
> GET /api/datasets/HuggingFaceTB/smoltalk2/parquet/Mid/Llama_Nemotron_Post_Training_Dataset_reasoning_r1/0.parquet HTTP/2
> Host: huggingface.co
> User-Agent: curl/8.7.1
> Accept: */*
> Range: bytes=217875070-218006142
>
* Request completely sent off
< HTTP/2 416
< content-type: text/html
< content-length: 49
< server: CloudFront
< date: Wed, 16 Jul 2025 14:58:41 GMT
< expires: Wed, 16 Jul 2025 14:58:41 GMT
< content-range: bytes */177
< x-cache: Error from cloudfront
< via: 1.1 e2f1bed2f82641d6d5439eac20a790ba.cloudfront.net (CloudFront)
< x-amz-cf-pop: MSP50-P1
< x-amz-cf-id: Mo8hn-EZLJqE_hoBday8DdhmVXhV3v9-Wg-EEHI6gX_fNlkanVIUBA==
<
{ [49 bytes data]
100 49 100 49 0 0 2215 0 --:--:-- --:--:-- --:--:-- 2227
* Connection #0 to host huggingface.co left intact
(.venv) Daft main** ≡❯ curl -v -L -H "Range: bytes=217875070-218006142" -o output.parquet "https://huggingface.co/api/datasets/HuggingFaceTB/smoltalk2/parquet/Mid/Llama_Nemotron_Post_Training_Dataset_reasoning_r1/0.parquet"
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0* Host huggingface.co:443 was resolved.
* IPv6: (none)
* IPv4: 18.160.102.96, 18.160.102.110, 18.160.102.4, 18.160.102.86
* Trying 18.160.102.96:443...
* Connected to huggingface.co (18.160.102.96) port 443
* ALPN: curl offers h2,http/1.1
* (304) (OUT), TLS handshake, Client hello (1):
} [319 bytes data]
* CAfile: /etc/ssl/cert.pem
* CApath: none
* (304) (IN), TLS handshake, Server hello (2):
{ [122 bytes data]
* (304) (IN), TLS handshake, Unknown (8):
{ [19 bytes data]
* (304) (IN), TLS handshake, Certificate (11):
{ [3821 bytes data]
* (304) (IN), TLS handshake, CERT verify (15):
{ [264 bytes data]
* (304) (IN), TLS handshake, Finished (20):
{ [36 bytes data]
* (304) (OUT), TLS handshake, Finished (20):
} [36 bytes data]
* SSL connection using TLSv1.3 / AEAD-AES128-GCM-SHA256 / [blank] / UNDEF
* ALPN: server accepted h2
* Server certificate:
* subject: CN=huggingface.co
* start date: Apr 13 00:00:00 2025 GMT
* expire date: May 12 23:59:59 2026 GMT
* subjectAltName: host "huggingface.co" matched cert's "huggingface.co"
* issuer: C=US; O=Amazon; CN=Amazon RSA 2048 M02
* SSL certificate verify ok.
* using HTTP/2
* [HTTP/2] [1] OPENED stream for https://huggingface.co/api/datasets/HuggingFaceTB/smoltalk2/parquet/Mid/Llama_Nemotron_Post_Training_Dataset_reasoning_r1/0.parquet
* [HTTP/2] [1] [:method: GET]
* [HTTP/2] [1] [:scheme: https]
* [HTTP/2] [1] [:authority: huggingface.co]
* [HTTP/2] [1] [:path: /api/datasets/HuggingFaceTB/smoltalk2/parquet/Mid/Llama_Nemotron_Post_Training_Dataset_reasoning_r1/0.parquet]
* [HTTP/2] [1] [user-agent: curl/8.7.1]
* [HTTP/2] [1] [accept: */*]
* [HTTP/2] [1] [range: bytes=217875070-218006142]
> GET /api/datasets/HuggingFaceTB/smoltalk2/parquet/Mid/Llama_Nemotron_Post_Training_Dataset_reasoning_r1/0.parquet HTTP/2
> Host: huggingface.co
> User-Agent: curl/8.7.1
> Accept: */*
> Range: bytes=217875070-218006142
>
* Request completely sent off
< HTTP/2 416
< content-type: text/html
< content-length: 49
< server: CloudFront
< date: Wed, 16 Jul 2025 14:58:42 GMT
< expires: Wed, 16 Jul 2025 14:58:42 GMT
< content-range: bytes */177
< x-cache: Error from cloudfront
< via: 1.1 bb352451e1eacf85f8786ee3ecd07eca.cloudfront.net (CloudFront)
< x-amz-cf-pop: MSP50-P1
< x-amz-cf-id: 9xy-CX9KvlS8Ye4eFr8jXMDobZHFkvdyvkLJGmK_qiwZQywCCwfq7Q==
<
{ [49 bytes data]
100 49 100 49 0 0 2381 0 --:--:-- --:--:-- --:--:-- 2450
* Connection #0 to host huggingface.co left intact
(.venv) Daft main** ≡❯ curl -v -L -H "Range: bytes=217875070-218006142" -o output.parquet "https://huggingface.co/api/datasets/HuggingFaceTB/smoltalk2/parquet/Mid/Llama_Nemotron_Post_Training_Dataset_reasoning_r1/0.parquet"
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0* Host huggingface.co:443 was resolved.
* IPv6: (none)
* IPv4: 18.160.102.96, 18.160.102.110, 18.160.102.4, 18.160.102.86
* Trying 18.160.102.96:443...
* Connected to huggingface.co (18.160.102.96) port 443
* ALPN: curl offers h2,http/1.1
* (304) (OUT), TLS handshake, Client hello (1):
} [319 bytes data]
* CAfile: /etc/ssl/cert.pem
* CApath: none
* (304) (IN), TLS handshake, Server hello (2):
{ [122 bytes data]
* (304) (IN), TLS handshake, Unknown (8):
{ [19 bytes data]
* (304) (IN), TLS handshake, Certificate (11):
{ [3821 bytes data]
* (304) (IN), TLS handshake, CERT verify (15):
{ [264 bytes data]
* (304) (IN), TLS handshake, Finished (20):
{ [36 bytes data]
* (304) (OUT), TLS handshake, Finished (20):
} [36 bytes data]
* SSL connection using TLSv1.3 / AEAD-AES128-GCM-SHA256 / [blank] / UNDEF
* ALPN: server accepted h2
* Server certificate:
* subject: CN=huggingface.co
* start date: Apr 13 00:00:00 2025 GMT
* expire date: May 12 23:59:59 2026 GMT
* subjectAltName: host "huggingface.co" matched cert's "huggingface.co"
* issuer: C=US; O=Amazon; CN=Amazon RSA 2048 M02
* SSL certificate verify ok.
* using HTTP/2
* [HTTP/2] [1] OPENED stream for https://huggingface.co/api/datasets/HuggingFaceTB/smoltalk2/parquet/Mid/Llama_Nemotron_Post_Training_Dataset_reasoning_r1/0.parquet
* [HTTP/2] [1] [:method: GET]
* [HTTP/2] [1] [:scheme: https]
* [HTTP/2] [1] [:authority: huggingface.co]
* [HTTP/2] [1] [:path: /api/datasets/HuggingFaceTB/smoltalk2/parquet/Mid/Llama_Nemotron_Post_Training_Dataset_reasoning_r1/0.parquet]
* [HTTP/2] [1] [user-agent: curl/8.7.1]
* [HTTP/2] [1] [accept: */*]
* [HTTP/2] [1] [range: bytes=217875070-218006142]
> GET /api/datasets/HuggingFaceTB/smoltalk2/parquet/Mid/Llama_Nemotron_Post_Training_Dataset_reasoning_r1/0.parquet HTTP/2
> Host: huggingface.co
> User-Agent: curl/8.7.1
> Accept: */*
> Range: bytes=217875070-218006142
>
* Request completely sent off
< HTTP/2 416
< content-type: text/html
< content-length: 49
< server: CloudFront
< date: Wed, 16 Jul 2025 14:58:43 GMT
< expires: Wed, 16 Jul 2025 14:58:43 GMT
< content-range: bytes */177
< x-cache: Error from cloudfront
< via: 1.1 873527676a354c5998cad133525df9c0.cloudfront.net (CloudFront)
< x-amz-cf-pop: MSP50-P1
< x-amz-cf-id: wtBgwY4u4YJ2pD1ovM8UV770UiJoqWfs7i7VzschDyoLv5g7swGGmw==
<
{ [49 bytes data]
100 49 100 49 0 0 2273 0 --:--:-- --:--:-- --:--:-- 2333
* Connection #0 to host huggingface.co left intact
(.venv) Daft main** ≡❯ curl -v -L -H "Range: bytes=217875070-218006142" -o output.parquet "https://huggingface.co/api/datasets/HuggingFaceTB/smoltalk2/parquet/Mid/Llama_Nemotron_Post_Training_Dataset_reasoning_r1/0.parquet"
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0* Host huggingface.co:443 was resolved.
* IPv6: (none)
* IPv4: 18.160.102.96, 18.160.102.110, 18.160.102.4, 18.160.102.86
* Trying 18.160.102.96:443...
* Connected to huggingface.co (18.160.102.96) port 443
* ALPN: curl offers h2,http/1.1
* (304) (OUT), TLS handshake, Client hello (1):
} [319 bytes data]
* CAfile: /etc/ssl/cert.pem
* CApath: none
* (304) (IN), TLS handshake, Server hello (2):
{ [122 bytes data]
* (304) (IN), TLS handshake, Unknown (8):
{ [19 bytes data]
* (304) (IN), TLS handshake, Certificate (11):
{ [3821 bytes data]
* (304) (IN), TLS handshake, CERT verify (15):
{ [264 bytes data]
* (304) (IN), TLS handshake, Finished (20):
{ [36 bytes data]
* (304) (OUT), TLS handshake, Finished (20):
} [36 bytes data]
* SSL connection using TLSv1.3 / AEAD-AES128-GCM-SHA256 / [blank] / UNDEF
* ALPN: server accepted h2
* Server certificate:
* subject: CN=huggingface.co
* start date: Apr 13 00:00:00 2025 GMT
* expire date: May 12 23:59:59 2026 GMT
* subjectAltName: host "huggingface.co" matched cert's "huggingface.co"
* issuer: C=US; O=Amazon; CN=Amazon RSA 2048 M02
* SSL certificate verify ok.
* using HTTP/2
* [HTTP/2] [1] OPENED stream for https://huggingface.co/api/datasets/HuggingFaceTB/smoltalk2/parquet/Mid/Llama_Nemotron_Post_Training_Dataset_reasoning_r1/0.parquet
* [HTTP/2] [1] [:method: GET]
* [HTTP/2] [1] [:scheme: https]
* [HTTP/2] [1] [:authority: huggingface.co]
* [HTTP/2] [1] [:path: /api/datasets/HuggingFaceTB/smoltalk2/parquet/Mid/Llama_Nemotron_Post_Training_Dataset_reasoning_r1/0.parquet]
* [HTTP/2] [1] [user-agent: curl/8.7.1]
* [HTTP/2] [1] [accept: */*]
* [HTTP/2] [1] [range: bytes=217875070-218006142]
> GET /api/datasets/HuggingFaceTB/smoltalk2/parquet/Mid/Llama_Nemotron_Post_Training_Dataset_reasoning_r1/0.parquet HTTP/2
> Host: huggingface.co
> User-Agent: curl/8.7.1
> Accept: */*
> Range: bytes=217875070-218006142
>
* Request completely sent off
< HTTP/2 302
< content-type: text/plain; charset=utf-8
< content-length: 177
< location: https://huggingface.co/datasets/HuggingFaceTB/smoltalk2/resolve/refs%2Fconvert%2Fparquet/Mid/Llama_Nemotron_Post_Training_Dataset_reasoning_r1/0000.parquet
< date: Wed, 16 Jul 2025 14:58:44 GMT
< x-powered-by: huggingface-moon
< cross-origin-opener-policy: same-origin
< referrer-policy: strict-origin-when-cross-origin
< x-request-id: Root=1-6877be24-476860f03849cb1a1570c9d8
< access-control-allow-origin: https://huggingface.co
< access-control-expose-headers: X-Repo-Commit,X-Request-Id,X-Error-Code,X-Error-Message,X-Total-Count,ETag,Link,Accept-Ranges,Content-Range,X-Linked-Size,X-Linked-ETag,X-Xet-Hash
< set-cookie: token=; Path=/; Expires=Thu, 01 Jan 1970 00:00:00 GMT; Secure; SameSite=None
< set-cookie: token=; Domain=huggingface.co; Path=/; Expires=Thu, 01 Jan 1970 00:00:00 GMT; Secure; SameSite=Lax
< x-cache: Miss from cloudfront
< via: 1.1 dd5af138aa8a11d8a70d5ef690ad1a2a.cloudfront.net (CloudFront)
< x-amz-cf-pop: MSP50-P1
< x-amz-cf-id: xuSi0X5RpH1OZqQOM8gGQLQLU8eOM6Gbkk-bgIX_qBnTTaa1VNkExA==
<
* Ignoring the response-body
100 177 100 177 0 0 2021 0 --:--:-- --:--:-- --:--:-- 2034
* Connection #0 to host huggingface.co left intact
* Issue another request to this URL: 'https://huggingface.co/datasets/HuggingFaceTB/smoltalk2/resolve/refs%2Fconvert%2Fparquet/Mid/Llama_Nemotron_Post_Training_Dataset_reasoning_r1/0000.parquet'
* Found bundle for host: 0x600002d54570 [can multiplex]
* Re-using existing connection with host huggingface.co
* [HTTP/2] [3] OPENED stream for https://huggingface.co/datasets/HuggingFaceTB/smoltalk2/resolve/refs%2Fconvert%2Fparquet/Mid/Llama_Nemotron_Post_Training_Dataset_reasoning_r1/0000.parquet
* [HTTP/2] [3] [:method: GET]
* [HTTP/2] [3] [:scheme: https]
* [HTTP/2] [3] [:authority: huggingface.co]
* [HTTP/2] [3] [:path: /datasets/HuggingFaceTB/smoltalk2/resolve/refs%2Fconvert%2Fparquet/Mid/Llama_Nemotron_Post_Training_Dataset_reasoning_r1/0000.parquet]
* [HTTP/2] [3] [user-agent: curl/8.7.1]
* [HTTP/2] [3] [accept: */*]
* [HTTP/2] [3] [range: bytes=217875070-218006142]
> GET /datasets/HuggingFaceTB/smoltalk2/resolve/refs%2Fconvert%2Fparquet/Mid/Llama_Nemotron_Post_Training_Dataset_reasoning_r1/0000.parquet HTTP/2
> Host: huggingface.co
> User-Agent: curl/8.7.1
> Accept: */*
> Range: bytes=217875070-218006142
>
* Request completely sent off
< HTTP/2 302
< content-type: text/plain; charset=utf-8
< content-length: 1317
< location: https://cas-bridge.xethub.hf.co/xet-bridge-us/686fc33898943c873b45c9a0/cf8a3a5665cf8b2ff667fb5236a1e5cb13c7582955f9533c88e1387997ef3af9?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=cas%2F20250716%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20250716T145416Z&X-Amz-Expires=3600&X-Amz-Signature=21a15b50740d73fd8ce82d5105733ca067d2e612ada22570e09e93ebcc7f8842&X-Amz-SignedHeaders=host&X-Xet-Cas-Uid=public&response-content-disposition=inline%3B+filename*%3DUTF-8%27%270000.parquet%3B+filename%3D%220000.parquet%22%3B&x-id=GetObject&Expires=1752681256&Policy=eyJTdGF0ZW1lbnQiOlt7IkNvbmRpdGlvbiI6eyJEYXRlTGVzc1RoYW4iOnsiQVdTOkVwb2NoVGltZSI6MTc1MjY4MTI1Nn19LCJSZXNvdXJjZSI6Imh0dHBzOi8vY2FzLWJyaWRnZS54ZXRodWIuaGYuY28veGV0LWJyaWRnZS11cy82ODZmYzMzODk4OTQzYzg3M2I0NWM5YTAvY2Y4YTNhNTY2NWNmOGIyZmY2NjdmYjUyMzZhMWU1Y2IxM2M3NTgyOTU1Zjk1MzNjODhlMTM4Nzk5N2VmM2FmOSoifV19&Signature=Tl3xorJ-7yaWvG6Y1AhhRlV2Wko9QpoK1tdPOfNZaRbHo%7EdaAkJRJfcLAYD5YzozfHWBZMLlJsaMPJ1MAne21nr5%7E737sE6yLfBwHdP3ZFZhgrLsN%7EvkIWK2GYX543qTg-pVsf3it92w1oWyoyYNQ9srxLfEIuG2AKV2Nu3Ejl7S%7EaAq4Gv4jNemvRTLBFGgYPdUeuavudl4OD4RGkSGTnpzh-P-OBk5WvgpdZZnbb1cRAP73tFHsPDX4%7ETfQIor109G%7E0TB3Jq0wopO9WV0sMQyQs9peZc6bxONiTxb9aHM4yNvWNbVGtlPuC6YS4c9T1e9%7EehdgU4sDOI%7EhpaCvg__&Key-Pair-Id=K2L8F4GPSG1IFC
< date: Wed, 16 Jul 2025 14:58:44 GMT
< x-powered-by: huggingface-moon
< cross-origin-opener-policy: same-origin
< referrer-policy: strict-origin-when-cross-origin
< x-request-id: Root=1-6877be24-4f628b292dc8a7a5339c41d3
< access-control-allow-origin: https://huggingface.co
< vary: Origin, Accept
< access-control-expose-headers: X-Repo-Commit,X-Request-Id,X-Error-Code,X-Error-Message,X-Total-Count,ETag,Link,Accept-Ranges,Content-Range,X-Linked-Size,X-Linked-ETag,X-Xet-Hash
< set-cookie: token=; Path=/; Expires=Thu, 01 Jan 1970 00:00:00 GMT; Secure; SameSite=None
< set-cookie: token=; Domain=huggingface.co; Path=/; Expires=Thu, 01 Jan 1970 00:00:00 GMT; Secure; SameSite=Lax
< x-repo-commit: 712df366ffbc959d9f4279bf2da579230b7ca5d8
< accept-ranges: bytes
< x-linked-size: 218006142
< x-linked-etag: "01736bf26d0046ddec4ab8900fba3f0dc6500b038314b44d0edb73a7c88dec07"
< x-xet-hash: cf8a3a5665cf8b2ff667fb5236a1e5cb13c7582955f9533c88e1387997ef3af9
< link: <https://huggingface.co/api/datasets/HuggingFaceTB/smoltalk2/xet-read-token/712df366ffbc959d9f4279bf2da579230b7ca5d8>; rel="xet-auth", <https://cas-server.xethub.hf.co/reconstruction/cf8a3a5665cf8b2ff667fb5236a1e5cb13c7582955f9533c88e1387997ef3af9>; rel="xet-reconstruction-info"
< x-cache: Miss from cloudfront
< via: 1.1 dd5af138aa8a11d8a70d5ef690ad1a2a.cloudfront.net (CloudFront)
< x-amz-cf-pop: MSP50-P1
< x-amz-cf-id: 0qXw2sJGrWCLVt7c-Vtn09uE3nu6CrJw9RmAKvNr_flG75muclvlIg==
<
* Ignoring the response-body
100 1317 100 1317 0 0 9268 0 --:--:-- --:--:-- --:--:-- 9268
* Connection #0 to host huggingface.co left intact
* Issue another request to this URL: 'https://cas-bridge.xethub.hf.co/xet-bridge-us/686fc33898943c873b45c9a0/cf8a3a5665cf8b2ff667fb5236a1e5cb13c7582955f9533c88e1387997ef3af9?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=cas%2F20250716%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20250716T145416Z&X-Amz-Expires=3600&X-Amz-Signature=21a15b50740d73fd8ce82d5105733ca067d2e612ada22570e09e93ebcc7f8842&X-Amz-SignedHeaders=host&X-Xet-Cas-Uid=public&response-content-disposition=inline%3B+filename*%3DUTF-8%27%270000.parquet%3B+filename%3D%220000.parquet%22%3B&x-id=GetObject&Expires=1752681256&Policy=eyJTdGF0ZW1lbnQiOlt7IkNvbmRpdGlvbiI6eyJEYXRlTGVzc1RoYW4iOnsiQVdTOkVwb2NoVGltZSI6MTc1MjY4MTI1Nn19LCJSZXNvdXJjZSI6Imh0dHBzOi8vY2FzLWJyaWRnZS54ZXRodWIuaGYuY28veGV0LWJyaWRnZS11cy82ODZmYzMzODk4OTQzYzg3M2I0NWM5YTAvY2Y4YTNhNTY2NWNmOGIyZmY2NjdmYjUyMzZhMWU1Y2IxM2M3NTgyOTU1Zjk1MzNjODhlMTM4Nzk5N2VmM2FmOSoifV19&Signature=Tl3xorJ-7yaWvG6Y1AhhRlV2Wko9QpoK1tdPOfNZaRbHo%7EdaAkJRJfcLAYD5YzozfHWBZMLlJsaMPJ1MAne21nr5%7E737sE6yLfBwHdP3ZFZhgrLsN%7EvkIWK2GYX543qTg-pVsf3it92w1oWyoyYNQ9srxLfEIuG2AKV2Nu3Ejl7S%7EaAq4Gv4jNemvRTLBFGgYPdUeuavudl4OD4RGkSGTnpzh-P-OBk5WvgpdZZnbb1cRAP73tFHsPDX4%7ETfQIor109G%7E0TB3Jq0wopO9WV0sMQyQs9peZc6bxONiTxb9aHM4yNvWNbVGtlPuC6YS4c9T1e9%7EehdgU4sDOI%7EhpaCvg__&Key-Pair-Id=K2L8F4GPSG1IFC'
* Host cas-bridge.xethub.hf.co:443 was resolved.
* IPv6: (none)
* IPv4: 18.160.181.55, 18.160.181.54, 18.160.181.52, 18.160.181.88
* Trying 18.160.181.55:443...
* Connected to cas-bridge.xethub.hf.co (18.160.181.55) port 443
* ALPN: curl offers h2,http/1.1
* (304) (OUT), TLS handshake, Client hello (1):
} [328 bytes data]
* (304) (IN), TLS handshake, Server hello (2):
{ [122 bytes data]
* (304) (IN), TLS handshake, Unknown (8):
{ [19 bytes data]
* (304) (IN), TLS handshake, Certificate (11):
{ [3818 bytes data]
* (304) (IN), TLS handshake, CERT verify (15):
{ [264 bytes data]
* (304) (IN), TLS handshake, Finished (20):
{ [36 bytes data]
* (304) (OUT), TLS handshake, Finished (20):
} [36 bytes data]
* SSL connection using TLSv1.3 / AEAD-AES128-GCM-SHA256 / [blank] / UNDEF
* ALPN: server accepted h2
* Server certificate:
* subject: CN=cas-bridge.xethub.hf.co
* start date: Jun 4 00:00:00 2025 GMT
* expire date: Jul 3 23:59:59 2026 GMT
* subjectAltName: host "cas-bridge.xethub.hf.co" matched cert's "cas-bridge.xethub.hf.co"
* issuer: C=US; O=Amazon; CN=Amazon RSA 2048 M04
* SSL certificate verify ok.
* using HTTP/2
* [HTTP/2] [1] OPENED stream for https://cas-bridge.xethub.hf.co/xet-bridge-us/686fc33898943c873b45c9a0/cf8a3a5665cf8b2ff667fb5236a1e5cb13c7582955f9533c88e1387997ef3af9?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=cas%2F20250716%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20250716T145416Z&X-Amz-Expires=3600&X-Amz-Signature=21a15b50740d73fd8ce82d5105733ca067d2e612ada22570e09e93ebcc7f8842&X-Amz-SignedHeaders=host&X-Xet-Cas-Uid=public&response-content-disposition=inline%3B+filename*%3DUTF-8%27%270000.parquet%3B+filename%3D%220000.parquet%22%3B&x-id=GetObject&Expires=1752681256&Policy=eyJTdGF0ZW1lbnQiOlt7IkNvbmRpdGlvbiI6eyJEYXRlTGVzc1RoYW4iOnsiQVdTOkVwb2NoVGltZSI6MTc1MjY4MTI1Nn19LCJSZXNvdXJjZSI6Imh0dHBzOi8vY2FzLWJyaWRnZS54ZXRodWIuaGYuY28veGV0LWJyaWRnZS11cy82ODZmYzMzODk4OTQzYzg3M2I0NWM5YTAvY2Y4YTNhNTY2NWNmOGIyZmY2NjdmYjUyMzZhMWU1Y2IxM2M3NTgyOTU1Zjk1MzNjODhlMTM4Nzk5N2VmM2FmOSoifV19&Signature=Tl3xorJ-7yaWvG6Y1AhhRlV2Wko9QpoK1tdPOfNZaRbHo%7EdaAkJRJfcLAYD5YzozfHWBZMLlJsaMPJ1MAne21nr5%7E737sE6yLfBwHdP3ZFZhgrLsN%7EvkIWK2GYX543qTg-pVsf3it92w1oWyoyYNQ9srxLfEIuG2AKV2Nu3Ejl7S%7EaAq4Gv4jNemvRTLBFGgYPdUeuavudl4OD4RGkSGTnpzh-P-OBk5WvgpdZZnbb1cRAP73tFHsPDX4%7ETfQIor109G%7E0TB3Jq0wopO9WV0sMQyQs9peZc6bxONiTxb9aHM4yNvWNbVGtlPuC6YS4c9T1e9%7EehdgU4sDOI%7EhpaCvg__&Key-Pair-Id=K2L8F4GPSG1IFC
* [HTTP/2] [1] [:method: GET]
* [HTTP/2] [1] [:scheme: https]
* [HTTP/2] [1] [:authority: cas-bridge.xethub.hf.co]
* [HTTP/2] [1] [:path: /xet-bridge-us/686fc33898943c873b45c9a0/cf8a3a5665cf8b2ff667fb5236a1e5cb13c7582955f9533c88e1387997ef3af9?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=cas%2F20250716%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20250716T145416Z&X-Amz-Expires=3600&X-Amz-Signature=21a15b50740d73fd8ce82d5105733ca067d2e612ada22570e09e93ebcc7f8842&X-Amz-SignedHeaders=host&X-Xet-Cas-Uid=public&response-content-disposition=inline%3B+filename*%3DUTF-8%27%270000.parquet%3B+filename%3D%220000.parquet%22%3B&x-id=GetObject&Expires=1752681256&Policy=eyJTdGF0ZW1lbnQiOlt7IkNvbmRpdGlvbiI6eyJEYXRlTGVzc1RoYW4iOnsiQVdTOkVwb2NoVGltZSI6MTc1MjY4MTI1Nn19LCJSZXNvdXJjZSI6Imh0dHBzOi8vY2FzLWJyaWRnZS54ZXRodWIuaGYuY28veGV0LWJyaWRnZS11cy82ODZmYzMzODk4OTQzYzg3M2I0NWM5YTAvY2Y4YTNhNTY2NWNmOGIyZmY2NjdmYjUyMzZhMWU1Y2IxM2M3NTgyOTU1Zjk1MzNjODhlMTM4Nzk5N2VmM2FmOSoifV19&Signature=Tl3xorJ-7yaWvG6Y1AhhRlV2Wko9QpoK1tdPOfNZaRbHo%7EdaAkJRJfcLAYD5YzozfHWBZMLlJsaMPJ1MAne21nr5%7E737sE6yLfBwHdP3ZFZhgrLsN%7EvkIWK2GYX543qTg-pVsf3it92w1oWyoyYNQ9srxLfEIuG2AKV2Nu3Ejl7S%7EaAq4Gv4jNemvRTLBFGgYPdUeuavudl4OD4RGkSGTnpzh-P-OBk5WvgpdZZnbb1cRAP73tFHsPDX4%7ETfQIor109G%7E0TB3Jq0wopO9WV0sMQyQs9peZc6bxONiTxb9aHM4yNvWNbVGtlPuC6YS4c9T1e9%7EehdgU4sDOI%7EhpaCvg__&Key-Pair-Id=K2L8F4GPSG1IFC]
* [HTTP/2] [1] [user-agent: curl/8.7.1]
* [HTTP/2] [1] [accept: */*]
* [HTTP/2] [1] [range: bytes=217875070-218006142]
> GET /xet-bridge-us/686fc33898943c873b45c9a0/cf8a3a5665cf8b2ff667fb5236a1e5cb13c7582955f9533c88e1387997ef3af9?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=cas%2F20250716%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20250716T145416Z&X-Amz-Expires=3600&X-Amz-Signature=21a15b50740d73fd8ce82d5105733ca067d2e612ada22570e09e93ebcc7f8842&X-Amz-SignedHeaders=host&X-Xet-Cas-Uid=public&response-content-disposition=inline%3B+filename*%3DUTF-8%27%270000.parquet%3B+filename%3D%220000.parquet%22%3B&x-id=GetObject&Expires=1752681256&Policy=eyJTdGF0ZW1lbnQiOlt7IkNvbmRpdGlvbiI6eyJEYXRlTGVzc1RoYW4iOnsiQVdTOkVwb2NoVGltZSI6MTc1MjY4MTI1Nn19LCJSZXNvdXJjZSI6Imh0dHBzOi8vY2FzLWJyaWRnZS54ZXRodWIuaGYuY28veGV0LWJyaWRnZS11cy82ODZmYzMzODk4OTQzYzg3M2I0NWM5YTAvY2Y4YTNhNTY2NWNmOGIyZmY2NjdmYjUyMzZhMWU1Y2IxM2M3NTgyOTU1Zjk1MzNjODhlMTM4Nzk5N2VmM2FmOSoifV19&Signature=Tl3xorJ-7yaWvG6Y1AhhRlV2Wko9QpoK1tdPOfNZaRbHo%7EdaAkJRJfcLAYD5YzozfHWBZMLlJsaMPJ1MAne21nr5%7E737sE6yLfBwHdP3ZFZhgrLsN%7EvkIWK2GYX543qTg-pVsf3it92w1oWyoyYNQ9srxLfEIuG2AKV2Nu3Ejl7S%7EaAq4Gv4jNemvRTLBFGgYPdUeuavudl4OD4RGkSGTnpzh-P-OBk5WvgpdZZnbb1cRAP73tFHsPDX4%7ETfQIor109G%7E0TB3Jq0wopO9WV0sMQyQs9peZc6bxONiTxb9aHM4yNvWNbVGtlPuC6YS4c9T1e9%7EehdgU4sDOI%7EhpaCvg__&Key-Pair-Id=K2L8F4GPSG1IFC HTTP/2
> Host: cas-bridge.xethub.hf.co
> User-Agent: curl/8.7.1
> Accept: */*
> Range: bytes=217875070-218006142
>
* Request completely sent off
< HTTP/2 206
< content-length: 131072
< date: Mon, 14 Jul 2025 08:40:28 GMT
< x-request-id: 01K041FDPVA03RR2PRXDZSN30G
< content-disposition: inline; filename*=UTF-8''0000.parquet; filename="0000.parquet";
< cache-control: public, max-age=31536000
< etag: "cf8a3a5665cf8b2ff667fb5236a1e5cb13c7582955f9533c88e1387997ef3af9"
< access-control-allow-origin: *
< access-control-allow-headers: Content-Range, Content-Type, Content-Disposition, ETag
< access-control-expose-headers: Accept-Ranges, Content-Range, Content-Type, Content-Disposition, ETag, X-Cache
< x-cache: Hit from cloudfront
< via: 1.1 1c857e24a4dc84d2d9c78d5b3463bed6.cloudfront.net (CloudFront)
< x-amz-cf-pop: MSP50-P2
< x-amz-cf-id: 3SxFmQa5wLeeXbNiwaAo0_RwoR_n7-SivjsLjDLG-Pwn5UhG2oiEQA==
< age: 195496
< content-security-policy: default-src 'none'; sandbox
< content-range: bytes 217875070-218006141/218006142
<
{ [8192 bytes data]
100 128k 100 128k 0 0 769k 0 --:--:-- --:--:-- --:--:-- 769k
* Connection #1 to host cas-bridge.xethub.hf.co left intact
</details>
### Expected behavior
always get back a `206`
### Environment info
n/a | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7685/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7685/timeline | null | null | null | null |
https://api.github.com/repos/huggingface/datasets/issues/7684 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7684/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7684/comments | https://api.github.com/repos/huggingface/datasets/issues/7684/events | https://github.com/huggingface/datasets/pull/7684 | 3,231,680,474 | PR_kwDODunzps6e9SjQ | 7,684 | fix audio cast storage from array + sampling_rate | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7684). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-07-15T10:13:42Z | 2025-07-15T10:24:08Z | 2025-07-15T10:24:07Z | MEMBER | null | null | null | fix https://github.com/huggingface/datasets/issues/7682 | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7684/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7684/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7684.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7684",
"merged_at": "2025-07-15T10:24:07Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7684.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7684"
} |
https://api.github.com/repos/huggingface/datasets/issues/7683 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7683/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7683/comments | https://api.github.com/repos/huggingface/datasets/issues/7683/events | https://github.com/huggingface/datasets/pull/7683 | 3,231,553,161 | PR_kwDODunzps6e82iW | 7,683 | Convert to string when needed + faster .zstd | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7683). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-07-15T09:37:44Z | 2025-07-15T10:13:58Z | 2025-07-15T10:13:56Z | MEMBER | null | null | null | for https://huggingface.co/datasets/allenai/olmo-mix-1124 | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7683/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7683/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7683.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7683",
"merged_at": "2025-07-15T10:13:56Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7683.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7683"
} |
https://api.github.com/repos/huggingface/datasets/issues/7682 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7682/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7682/comments | https://api.github.com/repos/huggingface/datasets/issues/7682/events | https://github.com/huggingface/datasets/issues/7682 | 3,229,687,253 | I_kwDODunzps7AgR3V | 7,682 | Fail to cast Audio feature for numpy arrays in datasets 4.0.0 | {
"avatar_url": "https://avatars.githubusercontent.com/u/163345686?v=4",
"events_url": "https://api.github.com/users/luatil-cloud/events{/privacy}",
"followers_url": "https://api.github.com/users/luatil-cloud/followers",
"following_url": "https://api.github.com/users/luatil-cloud/following{/other_user}",
"gists_url": "https://api.github.com/users/luatil-cloud/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/luatil-cloud",
"id": 163345686,
"login": "luatil-cloud",
"node_id": "U_kgDOCbx1Fg",
"organizations_url": "https://api.github.com/users/luatil-cloud/orgs",
"received_events_url": "https://api.github.com/users/luatil-cloud/received_events",
"repos_url": "https://api.github.com/users/luatil-cloud/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/luatil-cloud/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/luatil-cloud/subscriptions",
"type": "User",
"url": "https://api.github.com/users/luatil-cloud",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"thanks for reporting, I opened a PR and I'll make a patch release soon ",
"> thanks for reporting, I opened a PR and I'll make a patch release soon\n\nThank you very much @lhoestq!"
] | 2025-07-14T18:41:02Z | 2025-07-15T12:10:39Z | 2025-07-15T10:24:08Z | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
Casting features with Audio for numpy arrays - done here with `ds.map(gen_sine, features=features)` fails
in version 4.0.0 but not in version 3.6.0
### Steps to reproduce the bug
The following `uv script` should be able to reproduce the bug in version 4.0.0
and pass in version 3.6.0 on a macOS Sequoia 15.5
```python
# /// script
# requires-python = ">=3.13"
# dependencies = [
# "datasets[audio]==4.0.0",
# "librosa>=0.11.0",
# ]
# ///
# NAME
# create_audio_dataset.py - create an audio dataset of sine waves
#
# SYNOPSIS
# uv run create_audio_dataset.py
#
# DESCRIPTION
# Create an audio dataset using the Hugging Face [datasets] library.
# Illustrates how to create synthetic audio datasets using the [map]
# datasets function.
#
# The strategy is to first create a dataset with the input to the
# generation function, then execute the map function that generates
# the result, and finally cast the final features.
#
# BUG
# Casting features with Audio for numpy arrays -
# done here with `ds.map(gen_sine, features=features)` fails
# in version 4.0.0 but not in version 3.6.0
#
# This happens both in cases where --extra audio is provided and where is not.
# When audio is not provided i've installed the latest compatible version
# of soundfile.
#
# The error when soundfile is installed but the audio --extra is not
# indicates that the array values do not have the `.T` property,
# whilst also indicating that the value is a list instead of a numpy array.
#
# Last lines of error report when for datasets + soundfile case
# ...
#
# File "/Users/luasantilli/.cache/uv/archive-v0/tc_5IhQe7Zpw8ZXgQWpnl/lib/python3.13/site-packages/datasets/features/audio.py", line 239, in cast_storage
# storage = pa.array([Audio().encode_example(x) if x is not None else None for x in storage.to_pylist()])
# ~~~~~~~~~~~~~~~~~~~~~~^^^
# File "/Users/luasantilli/.cache/uv/archive-v0/tc_5IhQe7Zpw8ZXgQWpnl/lib/python3.13/site-packages/datasets/features/audio.py", line 122, in encode_example
# sf.write(buffer, value["array"].T, value["sampling_rate"], format="wav")
# ^^^^^^^^^^^^^^^^
# AttributeError: 'list' object has no attribute 'T'
# ...
#
# For the case of datasets[audio] without explicit adding soundfile I get an FFmpeg
# error.
#
# Last lines of error report:
#
# ...
# RuntimeError: Could not load libtorchcodec. Likely causes:
# 1. FFmpeg is not properly installed in your environment. We support
# versions 4, 5, 6 and 7.
# 2. The PyTorch version (2.7.1) is not compatible with
# this version of TorchCodec. Refer to the version compatibility
# table:
# https://github.com/pytorch/torchcodec?tab=readme-ov-file#installing-torchcodec.
# 3. Another runtime dependency; see exceptions below.
# The following exceptions were raised as we tried to load libtorchcodec:
#
# [start of libtorchcodec loading traceback]
# FFmpeg version 7: dlopen(/Users/luasantilli/.cache/uv/archive-v0/RK3IAlGfiICwDkHm2guLC/lib/python3.13/site-packages/torchcodec/libtorchcodec_decoder7.dylib, 0x0006): Library not loaded: @rpath/libavutil.59.dylib
# Referenced from: <6DB21246-F28A-31A6-910A-D8F3355D1064> /Users/luasantilli/.cache/uv/archive-v0/RK3IAlGfiICwDkHm2guLC/lib/python3.13/site-packages/torchcodec/libtorchcodec_decoder7.dylib
# Reason: no LC_RPATH's found
# FFmpeg version 6: dlopen(/Users/luasantilli/.cache/uv/archive-v0/RK3IAlGfiICwDkHm2guLC/lib/python3.13/site-packages/torchcodec/libtorchcodec_decoder6.dylib, 0x0006): Library not loaded: @rpath/libavutil.58.dylib
# Referenced from: <BD3B44FC-E14B-3ABF-800F-BB54B6CCA3B1> /Users/luasantilli/.cache/uv/archive-v0/RK3IAlGfiICwDkHm2guLC/lib/python3.13/site-packages/torchcodec/libtorchcodec_decoder6.dylib
# Reason: no LC_RPATH's found
# FFmpeg version 5: dlopen(/Users/luasantilli/.cache/uv/archive-v0/RK3IAlGfiICwDkHm2guLC/lib/python3.13/site-packages/torchcodec/libtorchcodec_decoder5.dylib, 0x0006): Library not loaded: @rpath/libavutil.57.dylib
# Referenced from: <F06EBF8A-238C-3A96-BFBB-B34E0BBDABF0> /Users/luasantilli/.cache/uv/archive-v0/RK3IAlGfiICwDkHm2guLC/lib/python3.13/site-packages/torchcodec/libtorchcodec_decoder5.dylib
# Reason: no LC_RPATH's found
# FFmpeg version 4: dlopen(/Users/luasantilli/.cache/uv/archive-v0/RK3IAlGfiICwDkHm2guLC/lib/python3.13/site-packages/torchcodec/libtorchcodec_decoder4.dylib, 0x0006): Library not loaded: @rpath/libavutil.56.dylib
# Referenced from: <6E59F017-C703-3AF6-A271-6277DD5F8170> /Users/luasantilli/.cache/uv/archive-v0/RK3IAlGfiICwDkHm2guLC/lib/python3.13/site-packages/torchcodec/libtorchcodec_decoder4.dylib
# Reason: no LC_RPATH's found
# ...
#
# This is strange because the the same error does not happen when using version 3.6.0 with datasets[audio].
#
# The same error appears in python3.12
## Imports
import numpy as np
from datasets import Dataset, Features, Audio, Value
## Parameters
NUM_WAVES = 128
SAMPLE_RATE = 16_000
DURATION = 1.0
## Input dataset arguments
freqs = np.linspace(100, 2000, NUM_WAVES).tolist()
ds = Dataset.from_dict({"frequency": freqs})
## Features for the final dataset
features = Features(
{"frequency": Value("float32"), "audio": Audio(sampling_rate=SAMPLE_RATE)}
)
## Generate audio sine waves and cast features
def gen_sine(example):
t = np.linspace(0, DURATION, int(SAMPLE_RATE * DURATION), endpoint=False)
wav = np.sin(2 * np.pi * example["frequency"] * t)
return {
"frequency": example["frequency"],
"audio": {"array": wav, "sampling_rate": SAMPLE_RATE},
}
ds = ds.map(gen_sine, features=features)
print(ds)
print(ds.features)
```
### Expected behavior
I expect the result of version `4.0.0` to be the same of that in version `3.6.0`. Switching the value
of the script above to `3.6.0` I get the following, expected, result:
```
$ uv run bug_report.py
Map: 100%|███████████████████████████████████████████████████████| 128/128 [00:00<00:00, 204.58 examples/s]
Dataset({
features: ['frequency', 'audio'],
num_rows: 128
})
{'frequency': Value(dtype='float32', id=None), 'audio': Audio(sampling_rate=16000, mono=True, decode=True, id=None)}
```
### Environment info
- `datasets` version: 4.0.0
- Platform: macOS-15.5-arm64-arm-64bit-Mach-O
- Python version: 3.13.1
- `huggingface_hub` version: 0.33.4
- PyArrow version: 20.0.0
- Pandas version: 2.3.1
- `fsspec` version: 2025.3.0 | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7682/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7682/timeline | null | completed | null | null |
https://api.github.com/repos/huggingface/datasets/issues/7681 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7681/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7681/comments | https://api.github.com/repos/huggingface/datasets/issues/7681/events | https://github.com/huggingface/datasets/issues/7681 | 3,227,112,736 | I_kwDODunzps7AWdUg | 7,681 | Probabilistic High Memory Usage and Freeze on Python 3.10 | {
"avatar_url": "https://avatars.githubusercontent.com/u/82735346?v=4",
"events_url": "https://api.github.com/users/ryan-minato/events{/privacy}",
"followers_url": "https://api.github.com/users/ryan-minato/followers",
"following_url": "https://api.github.com/users/ryan-minato/following{/other_user}",
"gists_url": "https://api.github.com/users/ryan-minato/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/ryan-minato",
"id": 82735346,
"login": "ryan-minato",
"node_id": "MDQ6VXNlcjgyNzM1MzQ2",
"organizations_url": "https://api.github.com/users/ryan-minato/orgs",
"received_events_url": "https://api.github.com/users/ryan-minato/received_events",
"repos_url": "https://api.github.com/users/ryan-minato/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/ryan-minato/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ryan-minato/subscriptions",
"type": "User",
"url": "https://api.github.com/users/ryan-minato",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [] | 2025-07-14T01:57:16Z | 2025-07-14T01:57:16Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
A probabilistic issue encountered when processing datasets containing PIL.Image columns using the huggingface/datasets library on Python 3.10. The process occasionally experiences a sudden and significant memory spike, reaching 100% utilization, leading to a complete freeze. During this freeze, the process becomes unresponsive, cannot be forcefully terminated, and does not throw any exceptions.
I have attempted to mitigate this issue by setting `datasets.config.IN_MEMORY_MAX_SIZE`, but it had no effect. In fact, based on the document of `load_dataset`, I suspect that setting `IN_MEMORY_MAX_SIZE` might even have a counterproductive effect.
This bug is not consistently reproducible, but its occurrence rate significantly decreases or disappears entirely when upgrading Python to version 3.11 or higher. Therefore, this issue also serves to share a potential solution for others who might encounter similar problems.
### Steps to reproduce the bug
Due to the probabilistic nature of this bug, consistent reproduction cannot be guaranteed for every run. However, in my environment, processing large datasets like timm/imagenet-1k-wds(whether reading, casting, or mapping operations) almost certainly triggers the issue at some point.
The probability of the issue occurring drastically increases when num_proc is set to a value greater than 1 during operations.
When the issue occurs, my system logs repeatedly show the following warnings:
```
WARN: very high memory utilization: 57.74GiB / 57.74GiB (100 %)
WARN: container is unhealthy: triggered memory limits (OOM)
WARN: container is unhealthy: triggered memory limits (OOM)
WARN: container is unhealthy: triggered memory limits (OOM)
```
### Expected behavior
The dataset should be read and processed normally without memory exhaustion or freezing. If an unrecoverable error occurs, an appropriate exception should be raised.
I have found that upgrading Python to version 3.11 or above completely resolves this issue. On Python 3.11, when memory usage approaches 100%, it suddenly drops before slowly increasing again. I suspect this behavior is due to an expected memory management action, possibly involving writing to disk cache, which prevents the complete freeze observed in Python 3.10.
### Environment info
- `datasets` version: 4.0.0
- Platform: Linux-5.15.0-71-generic-x86_64-with-glibc2.35
- Python version: 3.10.12
- `huggingface_hub` version: 0.33.4
- PyArrow version: 20.0.0
- Pandas version: 2.3.1
- `fsspec` version: 2025.3.0 | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7681/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7681/timeline | null | null | null | null |
https://api.github.com/repos/huggingface/datasets/issues/7680 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7680/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7680/comments | https://api.github.com/repos/huggingface/datasets/issues/7680/events | https://github.com/huggingface/datasets/issues/7680 | 3,224,824,151 | I_kwDODunzps7ANulX | 7,680 | Question about iterable dataset and streaming | {
"avatar_url": "https://avatars.githubusercontent.com/u/73541181?v=4",
"events_url": "https://api.github.com/users/Tavish9/events{/privacy}",
"followers_url": "https://api.github.com/users/Tavish9/followers",
"following_url": "https://api.github.com/users/Tavish9/following{/other_user}",
"gists_url": "https://api.github.com/users/Tavish9/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/Tavish9",
"id": 73541181,
"login": "Tavish9",
"node_id": "MDQ6VXNlcjczNTQxMTgx",
"organizations_url": "https://api.github.com/users/Tavish9/orgs",
"received_events_url": "https://api.github.com/users/Tavish9/received_events",
"repos_url": "https://api.github.com/users/Tavish9/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/Tavish9/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Tavish9/subscriptions",
"type": "User",
"url": "https://api.github.com/users/Tavish9",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"> If we have already loaded the dataset, why doing to_iterable_dataset? Does it go through the dataset faster than map-style dataset?\n\nyes, it makes a faster DataLoader for example (otherwise DataLoader uses `__getitem__` which is slower than iterating)\n\n> load_dataset(streaming=True) is useful for huge datase... | 2025-07-12T04:48:30Z | 2025-07-15T13:39:38Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | In the doc, I found the following example: https://github.com/huggingface/datasets/blob/611f5a592359ebac6f858f515c776aa7d99838b2/docs/source/stream.mdx?plain=1#L65-L78
I am confused,
1. If we have already loaded the dataset, why doing `to_iterable_dataset`? Does it go through the dataset faster than map-style dataset?
2. `load_dataset(streaming=True)` is useful for huge dataset, but the speed is slow. How to make it comparable to `to_iterable_dataset` without loading the whole dataset into RAM? | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7680/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7680/timeline | null | null | null | null |
https://api.github.com/repos/huggingface/datasets/issues/7679 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7679/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7679/comments | https://api.github.com/repos/huggingface/datasets/issues/7679/events | https://github.com/huggingface/datasets/issues/7679 | 3,220,787,371 | I_kwDODunzps6_-VCr | 7,679 | metric glue breaks with 4.0.0 | {
"avatar_url": "https://avatars.githubusercontent.com/u/10676103?v=4",
"events_url": "https://api.github.com/users/stas00/events{/privacy}",
"followers_url": "https://api.github.com/users/stas00/followers",
"following_url": "https://api.github.com/users/stas00/following{/other_user}",
"gists_url": "https://api.github.com/users/stas00/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/stas00",
"id": 10676103,
"login": "stas00",
"node_id": "MDQ6VXNlcjEwNjc2MTAz",
"organizations_url": "https://api.github.com/users/stas00/orgs",
"received_events_url": "https://api.github.com/users/stas00/received_events",
"repos_url": "https://api.github.com/users/stas00/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/stas00/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/stas00/subscriptions",
"type": "User",
"url": "https://api.github.com/users/stas00",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"I released `evaluate` 0.4.5 yesterday to fix the issue - sorry for the inconvenience:\n\n```\npip install -U evaluate\n```",
"Thanks so much, @lhoestq!"
] | 2025-07-10T21:39:50Z | 2025-07-11T17:42:01Z | 2025-07-11T17:42:01Z | CONTRIBUTOR | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
worked fine with 3.6.0, and with 4.0.0 `eval_metric = metric.compute()` in HF Accelerate breaks.
The code that fails is:
https://huggingface.co/spaces/evaluate-metric/glue/blob/v0.4.0/glue.py#L84
```
def simple_accuracy(preds, labels):
print(preds, labels)
print(f"{preds==labels}")
return float((preds == labels).mean())
```
data:
```
Column([1, 0, 0, 1, 1]) Column([1, 0, 0, 1, 0])
False
```
```
[rank0]: return float((preds == labels).mean())
[rank0]: ^^^^^^^^^^^^^^^^^^^^^^
[rank0]: AttributeError: 'bool' object has no attribute 'mean'
```
Some behavior has changed in this new major release of `datasets` and requires updating HF accelerate and perhaps the glue metric code, all belong to HF.
### Environment info
datasets=4.0.0 | {
"avatar_url": "https://avatars.githubusercontent.com/u/10676103?v=4",
"events_url": "https://api.github.com/users/stas00/events{/privacy}",
"followers_url": "https://api.github.com/users/stas00/followers",
"following_url": "https://api.github.com/users/stas00/following{/other_user}",
"gists_url": "https://api.github.com/users/stas00/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/stas00",
"id": 10676103,
"login": "stas00",
"node_id": "MDQ6VXNlcjEwNjc2MTAz",
"organizations_url": "https://api.github.com/users/stas00/orgs",
"received_events_url": "https://api.github.com/users/stas00/received_events",
"repos_url": "https://api.github.com/users/stas00/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/stas00/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/stas00/subscriptions",
"type": "User",
"url": "https://api.github.com/users/stas00",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7679/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7679/timeline | null | completed | null | null |
https://api.github.com/repos/huggingface/datasets/issues/7678 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7678/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7678/comments | https://api.github.com/repos/huggingface/datasets/issues/7678/events | https://github.com/huggingface/datasets/issues/7678 | 3,218,625,544 | I_kwDODunzps6_2FQI | 7,678 | To support decoding audio data, please install 'torchcodec'. | {
"avatar_url": "https://avatars.githubusercontent.com/u/48163702?v=4",
"events_url": "https://api.github.com/users/alpcansoydas/events{/privacy}",
"followers_url": "https://api.github.com/users/alpcansoydas/followers",
"following_url": "https://api.github.com/users/alpcansoydas/following{/other_user}",
"gists_url": "https://api.github.com/users/alpcansoydas/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/alpcansoydas",
"id": 48163702,
"login": "alpcansoydas",
"node_id": "MDQ6VXNlcjQ4MTYzNzAy",
"organizations_url": "https://api.github.com/users/alpcansoydas/orgs",
"received_events_url": "https://api.github.com/users/alpcansoydas/received_events",
"repos_url": "https://api.github.com/users/alpcansoydas/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/alpcansoydas/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/alpcansoydas/subscriptions",
"type": "User",
"url": "https://api.github.com/users/alpcansoydas",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"Hi ! yes you should `!pip install -U datasets[audio]` to have the required dependencies.\n\n`datasets` 4.0 now relies on `torchcodec` for audio decoding. The `torchcodec` AudioDecoder enables streaming from HF and also allows to decode ranges of audio"
] | 2025-07-10T09:43:13Z | 2025-07-11T05:05:42Z | 2025-07-11T05:05:42Z | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} |
In the latest version of datasets==4.0.0, i cannot print the audio data on the Colab notebook. But it works on the 3.6.0 version.
!pip install -q -U datasets huggingface_hub fsspec
from datasets import load_dataset
downloaded_dataset = load_dataset("ymoslem/MediaSpeech", "tr", split="train")
print(downloaded_dataset["audio"][0])
---------------------------------------------------------------------------
ImportError Traceback (most recent call last)
[/tmp/ipython-input-4-90623240.py](https://localhost:8080/#) in <cell line: 0>()
----> 1 downloaded_dataset["audio"][0]
10 frames
[/usr/local/lib/python3.11/dist-packages/datasets/features/audio.py](https://localhost:8080/#) in decode_example(self, value, token_per_repo_id)
170 from ._torchcodec import AudioDecoder
171 else:
--> 172 raise ImportError("To support decoding audio data, please install 'torchcodec'.")
173
174 if not self.decode:
ImportError: To support decoding audio data, please install 'torchcodec'.
### Environment info
- `datasets` version: 4.0.0
- Platform: Linux-6.1.123+-x86_64-with-glibc2.35
- Python version: 3.11.13
- `huggingface_hub` version: 0.33.2
- PyArrow version: 18.1.0
- Pandas version: 2.2.2
- `fsspec` version: 2025.3.0 | {
"avatar_url": "https://avatars.githubusercontent.com/u/48163702?v=4",
"events_url": "https://api.github.com/users/alpcansoydas/events{/privacy}",
"followers_url": "https://api.github.com/users/alpcansoydas/followers",
"following_url": "https://api.github.com/users/alpcansoydas/following{/other_user}",
"gists_url": "https://api.github.com/users/alpcansoydas/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/alpcansoydas",
"id": 48163702,
"login": "alpcansoydas",
"node_id": "MDQ6VXNlcjQ4MTYzNzAy",
"organizations_url": "https://api.github.com/users/alpcansoydas/orgs",
"received_events_url": "https://api.github.com/users/alpcansoydas/received_events",
"repos_url": "https://api.github.com/users/alpcansoydas/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/alpcansoydas/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/alpcansoydas/subscriptions",
"type": "User",
"url": "https://api.github.com/users/alpcansoydas",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7678/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7678/timeline | null | completed | null | null |
https://api.github.com/repos/huggingface/datasets/issues/7677 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7677/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7677/comments | https://api.github.com/repos/huggingface/datasets/issues/7677/events | https://github.com/huggingface/datasets/issues/7677 | 3,218,044,656 | I_kwDODunzps6_z3bw | 7,677 | Toxicity fails with datasets 4.0.0 | {
"avatar_url": "https://avatars.githubusercontent.com/u/82044803?v=4",
"events_url": "https://api.github.com/users/serena-ruan/events{/privacy}",
"followers_url": "https://api.github.com/users/serena-ruan/followers",
"following_url": "https://api.github.com/users/serena-ruan/following{/other_user}",
"gists_url": "https://api.github.com/users/serena-ruan/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/serena-ruan",
"id": 82044803,
"login": "serena-ruan",
"node_id": "MDQ6VXNlcjgyMDQ0ODAz",
"organizations_url": "https://api.github.com/users/serena-ruan/orgs",
"received_events_url": "https://api.github.com/users/serena-ruan/received_events",
"repos_url": "https://api.github.com/users/serena-ruan/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/serena-ruan/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/serena-ruan/subscriptions",
"type": "User",
"url": "https://api.github.com/users/serena-ruan",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"Hi ! You can fix this by upgrading `evaluate`:\n\n```\npip install -U evaluate\n```",
"Thanks, verified evaluate 0.4.5 works!"
] | 2025-07-10T06:15:22Z | 2025-07-11T04:40:59Z | 2025-07-11T04:40:59Z | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
With the latest 4.0.0 release, huggingface toxicity evaluation module fails with error: `ValueError: text input must be of type `str` (single example), `List[str]` (batch or single pretokenized example) or `List[List[str]]` (batch of pretokenized examples).`
### Steps to reproduce the bug
Repro:
```
>>> toxicity.compute(predictions=["This is a response"])
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/Users/serena.ruan/miniconda3/envs/mlflow-310/lib/python3.10/site-packages/evaluate/module.py", line 467, in compute
output = self._compute(**inputs, **compute_kwargs)
File "/Users/serena.ruan/.cache/huggingface/modules/evaluate_modules/metrics/evaluate-measurement--toxicity/2390290fa0bf6d78480143547c6b08f3d4f8805b249df8c7a8e80d0ce8e3778b/toxicity.py", line 135, in _compute
scores = toxicity(predictions, self.toxic_classifier, toxic_label)
File "/Users/serena.ruan/.cache/huggingface/modules/evaluate_modules/metrics/evaluate-measurement--toxicity/2390290fa0bf6d78480143547c6b08f3d4f8805b249df8c7a8e80d0ce8e3778b/toxicity.py", line 103, in toxicity
for pred_toxic in toxic_classifier(preds):
File "/Users/serena.ruan/miniconda3/envs/mlflow-310/lib/python3.10/site-packages/transformers/pipelines/text_classification.py", line 159, in __call__
result = super().__call__(*inputs, **kwargs)
File "/Users/serena.ruan/miniconda3/envs/mlflow-310/lib/python3.10/site-packages/transformers/pipelines/base.py", line 1431, in __call__
return self.run_single(inputs, preprocess_params, forward_params, postprocess_params)
File "/Users/serena.ruan/miniconda3/envs/mlflow-310/lib/python3.10/site-packages/transformers/pipelines/base.py", line 1437, in run_single
model_inputs = self.preprocess(inputs, **preprocess_params)
File "/Users/serena.ruan/miniconda3/envs/mlflow-310/lib/python3.10/site-packages/transformers/pipelines/text_classification.py", line 183, in preprocess
return self.tokenizer(inputs, return_tensors=return_tensors, **tokenizer_kwargs)
File "/Users/serena.ruan/miniconda3/envs/mlflow-310/lib/python3.10/site-packages/transformers/tokenization_utils_base.py", line 2867, in __call__
encodings = self._call_one(text=text, text_pair=text_pair, **all_kwargs)
File "/Users/serena.ruan/miniconda3/envs/mlflow-310/lib/python3.10/site-packages/transformers/tokenization_utils_base.py", line 2927, in _call_one
raise ValueError(
ValueError: text input must be of type `str` (single example), `List[str]` (batch or single pretokenized example) or `List[List[str]]` (batch of pretokenized examples).
```
### Expected behavior
This works before 4.0.0 release
### Environment info
- `datasets` version: 4.0.0
- Platform: macOS-15.5-arm64-arm-64bit
- Python version: 3.10.16
- `huggingface_hub` version: 0.33.0
- PyArrow version: 19.0.1
- Pandas version: 2.2.3
- `fsspec` version: 2024.12.0 | {
"avatar_url": "https://avatars.githubusercontent.com/u/82044803?v=4",
"events_url": "https://api.github.com/users/serena-ruan/events{/privacy}",
"followers_url": "https://api.github.com/users/serena-ruan/followers",
"following_url": "https://api.github.com/users/serena-ruan/following{/other_user}",
"gists_url": "https://api.github.com/users/serena-ruan/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/serena-ruan",
"id": 82044803,
"login": "serena-ruan",
"node_id": "MDQ6VXNlcjgyMDQ0ODAz",
"organizations_url": "https://api.github.com/users/serena-ruan/orgs",
"received_events_url": "https://api.github.com/users/serena-ruan/received_events",
"repos_url": "https://api.github.com/users/serena-ruan/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/serena-ruan/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/serena-ruan/subscriptions",
"type": "User",
"url": "https://api.github.com/users/serena-ruan",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7677/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7677/timeline | null | completed | null | null |
https://api.github.com/repos/huggingface/datasets/issues/7676 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7676/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7676/comments | https://api.github.com/repos/huggingface/datasets/issues/7676/events | https://github.com/huggingface/datasets/issues/7676 | 3,216,857,559 | I_kwDODunzps6_vVnX | 7,676 | Many things broken since the new 4.0.0 release | {
"avatar_url": "https://avatars.githubusercontent.com/u/37179323?v=4",
"events_url": "https://api.github.com/users/mobicham/events{/privacy}",
"followers_url": "https://api.github.com/users/mobicham/followers",
"following_url": "https://api.github.com/users/mobicham/following{/other_user}",
"gists_url": "https://api.github.com/users/mobicham/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/mobicham",
"id": 37179323,
"login": "mobicham",
"node_id": "MDQ6VXNlcjM3MTc5MzIz",
"organizations_url": "https://api.github.com/users/mobicham/orgs",
"received_events_url": "https://api.github.com/users/mobicham/received_events",
"repos_url": "https://api.github.com/users/mobicham/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/mobicham/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mobicham/subscriptions",
"type": "User",
"url": "https://api.github.com/users/mobicham",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"Happy to take a look, do you have a list of impacted datasets ?",
"Thanks @lhoestq , related to lm-eval, at least `winogrande`, `mmlu` and `hellaswag`, based on my tests yesterday. But many others like <a href=\"https://huggingface.co/datasets/lukaemon/bbh\">bbh</a>, most probably others too. ",
"Hi @mobicham ... | 2025-07-09T18:59:50Z | 2025-07-13T15:29:39Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
The new changes in 4.0.0 are breaking many datasets, including those from lm-evaluation-harness.
I am trying to revert back to older versions, like 3.6.0 to make the eval work but I keep getting:
``` Python
File /venv/main/lib/python3.12/site-packages/datasets/features/features.py:1474, in generate_from_dict(obj)
1471 class_type = _FEATURE_TYPES.get(_type, None) or globals().get(_type, None)
1473 if class_type is None:
-> 1474 raise ValueError(f"Feature type '{_type}' not found. Available feature types: {list(_FEATURE_TYPES.keys())}")
1476 if class_type == LargeList:
1477 feature = obj.pop("feature")
ValueError: Feature type 'List' not found. Available feature types: ['Value', 'ClassLabel', 'Translation', 'TranslationVariableLanguages', 'LargeList', 'Sequence', 'Array2D', 'Array3D', 'Array4D', 'Array5D', 'Audio', 'Image', 'Video', 'Pdf']
```
### Steps to reproduce the bug
``` Python
import lm_eval
model_eval = lm_eval.models.huggingface.HFLM(pretrained=model, tokenizer=tokenizer)
lm_eval.evaluator.simple_evaluate(model_eval, tasks=["winogrande"], num_fewshot=5, batch_size=1)
```
### Expected behavior
Older `datasets` versions should work just fine as before
### Environment info
- `datasets` version: 3.6.0
- Platform: Linux-6.8.0-60-generic-x86_64-with-glibc2.39
- Python version: 3.12.11
- `huggingface_hub` version: 0.33.1
- PyArrow version: 20.0.0
- Pandas version: 2.3.1
- `fsspec` version: 2025.3.0
| null | {
"+1": 11,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 11,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7676/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7676/timeline | null | null | null | null |
https://api.github.com/repos/huggingface/datasets/issues/7675 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7675/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7675/comments | https://api.github.com/repos/huggingface/datasets/issues/7675/events | https://github.com/huggingface/datasets/issues/7675 | 3,216,699,094 | I_kwDODunzps6_uu7W | 7,675 | common_voice_11_0.py failure in dataset library | {
"avatar_url": "https://avatars.githubusercontent.com/u/98793855?v=4",
"events_url": "https://api.github.com/users/egegurel/events{/privacy}",
"followers_url": "https://api.github.com/users/egegurel/followers",
"following_url": "https://api.github.com/users/egegurel/following{/other_user}",
"gists_url": "https://api.github.com/users/egegurel/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/egegurel",
"id": 98793855,
"login": "egegurel",
"node_id": "U_kgDOBeN5fw",
"organizations_url": "https://api.github.com/users/egegurel/orgs",
"received_events_url": "https://api.github.com/users/egegurel/received_events",
"repos_url": "https://api.github.com/users/egegurel/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/egegurel/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/egegurel/subscriptions",
"type": "User",
"url": "https://api.github.com/users/egegurel",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"Hi ! This dataset is not in a supported format and `datasets` 4 doesn't support datasets that based on python scripts which are often source of errors. Feel free to ask the dataset authors to convert the dataset to a supported format at https://huggingface.co/datasets/mozilla-foundation/common_voice_11_0/discussio... | 2025-07-09T17:47:59Z | 2025-07-15T18:25:19Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
I tried to download dataset but have got this error:
from datasets import load_dataset
load_dataset("mozilla-foundation/common_voice_11_0", "en", split="test", streaming=True)
---------------------------------------------------------------------------
RuntimeError Traceback (most recent call last)
Cell In[8], line 4
1 from datasets import load_dataset
----> 4 load_dataset("mozilla-foundation/common_voice_11_0", "en", split="test", streaming=True)
File c:\Users\ege_g\AppData\Local\Programs\Python\Python312\Lib\site-packages\datasets\load.py:1392, in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, verification_mode, keep_in_memory, save_infos, revision, token, streaming, num_proc, storage_options, **config_kwargs)
1387 verification_mode = VerificationMode(
1388 (verification_mode or VerificationMode.BASIC_CHECKS) if not save_infos else VerificationMode.ALL_CHECKS
1389 )
1391 # Create a dataset builder
-> 1392 builder_instance = load_dataset_builder(
1393 path=path,
1394 name=name,
1395 data_dir=data_dir,
1396 data_files=data_files,
1397 cache_dir=cache_dir,
1398 features=features,
1399 download_config=download_config,
1400 download_mode=download_mode,
1401 revision=revision,
1402 token=token,
1403 storage_options=storage_options,
1404 **config_kwargs,
1405 )
1407 # Return iterable dataset in case of streaming
1408 if streaming:
File c:\Users\ege_g\AppData\Local\Programs\Python\Python312\Lib\site-packages\datasets\load.py:1132, in load_dataset_builder(path, name, data_dir, data_files, cache_dir, features, download_config, download_mode, revision, token, storage_options, **config_kwargs)
1130 if features is not None:
1131 features = _fix_for_backward_compatible_features(features)
-> 1132 dataset_module = dataset_module_factory(
1133 path,
1134 revision=revision,
1135 download_config=download_config,
1136 download_mode=download_mode,
1137 data_dir=data_dir,
1138 data_files=data_files,
1139 cache_dir=cache_dir,
1140 )
1141 # Get dataset builder class
1142 builder_kwargs = dataset_module.builder_kwargs
File c:\Users\ege_g\AppData\Local\Programs\Python\Python312\Lib\site-packages\datasets\load.py:1031, in dataset_module_factory(path, revision, download_config, download_mode, data_dir, data_files, cache_dir, **download_kwargs)
1026 if isinstance(e1, FileNotFoundError):
1027 raise FileNotFoundError(
1028 f"Couldn't find any data file at {relative_to_absolute_path(path)}. "
1029 f"Couldn't find '{path}' on the Hugging Face Hub either: {type(e1).__name__}: {e1}"
1030 ) from None
-> 1031 raise e1 from None
1032 else:
1033 raise FileNotFoundError(f"Couldn't find any data file at {relative_to_absolute_path(path)}.")
File c:\Users\ege_g\AppData\Local\Programs\Python\Python312\Lib\site-packages\datasets\load.py:989, in dataset_module_factory(path, revision, download_config, download_mode, data_dir, data_files, cache_dir, **download_kwargs)
981 try:
982 api.hf_hub_download(
983 repo_id=path,
984 filename=filename,
(...)
987 proxies=download_config.proxies,
988 )
--> 989 raise RuntimeError(f"Dataset scripts are no longer supported, but found {filename}")
990 except EntryNotFoundError:
991 # Use the infos from the parquet export except in some cases:
992 if data_dir or data_files or (revision and revision != "main"):
RuntimeError: Dataset scripts are no longer supported, but found common_voice_11_0.py
### Steps to reproduce the bug
from datasets import load_dataset
load_dataset("mozilla-foundation/common_voice_11_0", "en", split="test", streaming=True)
### Expected behavior
its supposed to download this dataset.
### Environment info
Python 3.12 , Windows 11 | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7675/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7675/timeline | null | null | null | null |
https://api.github.com/repos/huggingface/datasets/issues/7674 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7674/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7674/comments | https://api.github.com/repos/huggingface/datasets/issues/7674/events | https://github.com/huggingface/datasets/pull/7674 | 3,216,251,069 | PR_kwDODunzps6eJGo5 | 7,674 | set dev version | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7674). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-07-09T15:01:25Z | 2025-07-09T15:04:01Z | 2025-07-09T15:01:33Z | MEMBER | null | null | null | null | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7674/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7674/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7674.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7674",
"merged_at": "2025-07-09T15:01:33Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7674.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7674"
} |
https://api.github.com/repos/huggingface/datasets/issues/7673 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7673/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7673/comments | https://api.github.com/repos/huggingface/datasets/issues/7673/events | https://github.com/huggingface/datasets/pull/7673 | 3,216,075,633 | PR_kwDODunzps6eIgj- | 7,673 | Release: 4.0.0 | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7673). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-07-09T14:03:16Z | 2025-07-09T14:36:19Z | 2025-07-09T14:36:18Z | MEMBER | null | null | null | null | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7673/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7673/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7673.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7673",
"merged_at": "2025-07-09T14:36:18Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7673.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7673"
} |
https://api.github.com/repos/huggingface/datasets/issues/7672 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7672/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7672/comments | https://api.github.com/repos/huggingface/datasets/issues/7672/events | https://github.com/huggingface/datasets/pull/7672 | 3,215,287,164 | PR_kwDODunzps6eF1vj | 7,672 | Fix double sequence | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7672). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-07-09T09:53:39Z | 2025-07-09T09:56:29Z | 2025-07-09T09:56:28Z | MEMBER | null | null | null | ```python
>>> Features({"a": Sequence(Sequence({"c": Value("int64")}))})
{'a': List({'c': List(Value('int64'))})}
```
instead of `{'a': {'c': List(List(Value('int64')))}}` | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7672/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7672/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7672.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7672",
"merged_at": "2025-07-09T09:56:27Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7672.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7672"
} |
https://api.github.com/repos/huggingface/datasets/issues/7671 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7671/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7671/comments | https://api.github.com/repos/huggingface/datasets/issues/7671/events | https://github.com/huggingface/datasets/issues/7671 | 3,213,223,886 | I_kwDODunzps6_hefO | 7,671 | Mapping function not working if the first example is returned as None | {
"avatar_url": "https://avatars.githubusercontent.com/u/46325823?v=4",
"events_url": "https://api.github.com/users/dnaihao/events{/privacy}",
"followers_url": "https://api.github.com/users/dnaihao/followers",
"following_url": "https://api.github.com/users/dnaihao/following{/other_user}",
"gists_url": "https://api.github.com/users/dnaihao/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/dnaihao",
"id": 46325823,
"login": "dnaihao",
"node_id": "MDQ6VXNlcjQ2MzI1ODIz",
"organizations_url": "https://api.github.com/users/dnaihao/orgs",
"received_events_url": "https://api.github.com/users/dnaihao/received_events",
"repos_url": "https://api.github.com/users/dnaihao/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/dnaihao/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dnaihao/subscriptions",
"type": "User",
"url": "https://api.github.com/users/dnaihao",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"Hi, map() always expect an output.\n\nIf you wish to filter examples, you should use filter(), in your case it could be something like this:\n\n```python\nds = ds.map(my_processing_function).filter(ignore_long_prompts)\n```",
"Realized this! Thanks a lot, I will close this issue then."
] | 2025-07-08T17:07:47Z | 2025-07-09T12:30:32Z | 2025-07-09T12:30:32Z | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
https://github.com/huggingface/datasets/blob/8a19de052e3d79f79cea26821454bbcf0e9dcd68/src/datasets/arrow_dataset.py#L3652C29-L3652C37
Here we can see the writer is initialized on `i==0`. However, there can be cases where in the user mapping function, the first example is filtered out (length constraints, etc).
In this case, the writer would be a `None` type and the code will report `NoneType has no write function`.
A simple fix is available, simply change line 3652 from `if i == 0:` to `if writer is None:`
### Steps to reproduce the bug
Prepare a dataset
have this function
```
import datasets
def make_map_fn(split, max_prompt_tokens=3):
def process_fn(example, idx):
question = example['question']
reasoning_steps = example['reasoning_steps']
label = example['label']
answer_format = ""
for i in range(len(reasoning_steps)):
system_message = "Dummy"
all_steps_formatted = []
content = f"""Dummy"""
prompt = [
{"role": "system", "content": system_message},
{"role": "user", "content": content},
]
tokenized = tokenizer.apply_chat_template(prompt, return_tensors="pt", truncation=False)
if tokenized.shape[1] > max_prompt_tokens:
return None # skip overly long examples
data = {
"dummy": "dummy"
}
return data
return process_fn
...
# load your dataset
...
train = train.map(function=make_map_fn('train'), with_indices=True)
```
### Expected behavior
The dataset mapping shall behave even when the first example is filtered out.
### Environment info
I am using `datasets==3.6.0` but I have observed this issue in the github repo too: https://github.com/huggingface/datasets/blob/8a19de052e3d79f79cea26821454bbcf0e9dcd68/src/datasets/arrow_dataset.py#L3652C29-L3652C37 | {
"avatar_url": "https://avatars.githubusercontent.com/u/46325823?v=4",
"events_url": "https://api.github.com/users/dnaihao/events{/privacy}",
"followers_url": "https://api.github.com/users/dnaihao/followers",
"following_url": "https://api.github.com/users/dnaihao/following{/other_user}",
"gists_url": "https://api.github.com/users/dnaihao/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/dnaihao",
"id": 46325823,
"login": "dnaihao",
"node_id": "MDQ6VXNlcjQ2MzI1ODIz",
"organizations_url": "https://api.github.com/users/dnaihao/orgs",
"received_events_url": "https://api.github.com/users/dnaihao/received_events",
"repos_url": "https://api.github.com/users/dnaihao/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/dnaihao/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dnaihao/subscriptions",
"type": "User",
"url": "https://api.github.com/users/dnaihao",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7671/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7671/timeline | null | completed | null | null |
https://api.github.com/repos/huggingface/datasets/issues/7670 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7670/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7670/comments | https://api.github.com/repos/huggingface/datasets/issues/7670/events | https://github.com/huggingface/datasets/pull/7670 | 3,208,962,372 | PR_kwDODunzps6dwgOc | 7,670 | Fix audio bytes | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7670). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-07-07T13:05:15Z | 2025-07-07T13:07:47Z | 2025-07-07T13:05:33Z | MEMBER | null | null | null | null | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7670/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7670/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7670.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7670",
"merged_at": "2025-07-07T13:05:33Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7670.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7670"
} |
https://api.github.com/repos/huggingface/datasets/issues/7669 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7669/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7669/comments | https://api.github.com/repos/huggingface/datasets/issues/7669/events | https://github.com/huggingface/datasets/issues/7669 | 3,203,541,091 | I_kwDODunzps6-8ihj | 7,669 | How can I add my custom data to huggingface datasets | {
"avatar_url": "https://avatars.githubusercontent.com/u/219205504?v=4",
"events_url": "https://api.github.com/users/xiagod/events{/privacy}",
"followers_url": "https://api.github.com/users/xiagod/followers",
"following_url": "https://api.github.com/users/xiagod/following{/other_user}",
"gists_url": "https://api.github.com/users/xiagod/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/xiagod",
"id": 219205504,
"login": "xiagod",
"node_id": "U_kgDODRDPgA",
"organizations_url": "https://api.github.com/users/xiagod/orgs",
"received_events_url": "https://api.github.com/users/xiagod/received_events",
"repos_url": "https://api.github.com/users/xiagod/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/xiagod/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/xiagod/subscriptions",
"type": "User",
"url": "https://api.github.com/users/xiagod",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"Hey @xiagod \n\nThe easiest way to add your custom data to Hugging Face Datasets is to use the built-in load_dataset function with your local files. Some examples include:\n\nCSV files:\nfrom datasets import load_dataset\ndataset = load_dataset(\"csv\", data_files=\"my_file.csv\")\n\nJSON or JSONL files:\nfrom dat... | 2025-07-04T19:19:54Z | 2025-07-05T18:19:37Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | I want to add my custom dataset in huggingface dataset. Please guide me how to achieve that. | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7669/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7669/timeline | null | null | null | null |
https://api.github.com/repos/huggingface/datasets/issues/7668 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7668/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7668/comments | https://api.github.com/repos/huggingface/datasets/issues/7668/events | https://github.com/huggingface/datasets/issues/7668 | 3,199,039,322 | I_kwDODunzps6-rXda | 7,668 | Broken EXIF crash the whole program | {
"avatar_url": "https://avatars.githubusercontent.com/u/30485844?v=4",
"events_url": "https://api.github.com/users/Seas0/events{/privacy}",
"followers_url": "https://api.github.com/users/Seas0/followers",
"following_url": "https://api.github.com/users/Seas0/following{/other_user}",
"gists_url": "https://api.github.com/users/Seas0/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/Seas0",
"id": 30485844,
"login": "Seas0",
"node_id": "MDQ6VXNlcjMwNDg1ODQ0",
"organizations_url": "https://api.github.com/users/Seas0/orgs",
"received_events_url": "https://api.github.com/users/Seas0/received_events",
"repos_url": "https://api.github.com/users/Seas0/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/Seas0/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Seas0/subscriptions",
"type": "User",
"url": "https://api.github.com/users/Seas0",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"There are other discussions about error handling for images decoding here : https://github.com/huggingface/datasets/issues/7632 https://github.com/huggingface/datasets/issues/7612\n\nand a PR here: https://github.com/huggingface/datasets/pull/7638 (would love your input on the proposed solution !)"
] | 2025-07-03T11:24:15Z | 2025-07-03T12:27:16Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
When parsing this image in the ImageNet1K dataset, the `datasets` crashs whole training process just because unable to parse an invalid EXIF tag.

### Steps to reproduce the bug
Use the `datasets.Image.decode_example` method to decode the aforementioned image could reproduce the bug.
The decoding function will throw an unhandled exception at the `image.getexif()` method call due to invalid utf-8 stream in EXIF tags.
```
File lib/python3.12/site-packages/datasets/features/image.py:188, in Image.decode_example(self, value, token_per_repo_id)
186 image = PIL.Image.open(BytesIO(bytes_))
187 image.load() # to avoid "Too many open files" errors
--> 188 if image.getexif().get(PIL.Image.ExifTags.Base.Orientation) is not None:
189 image = PIL.ImageOps.exif_transpose(image)
190 if self.mode and self.mode != image.mode:
File lib/python3.12/site-packages/PIL/Image.py:1542, in Image.getexif(self)
1540 xmp_tags = self.info.get("XML:com.adobe.xmp")
1541 if not xmp_tags and (xmp_tags := self.info.get("xmp")):
-> 1542 xmp_tags = xmp_tags.decode("utf-8")
1543 if xmp_tags:
1544 match = re.search(r'tiff:Orientation(="|>)([0-9])', xmp_tags)
UnicodeDecodeError: 'utf-8' codec can't decode byte 0xa8 in position 4312: invalid start byte
```
### Expected behavior
The invalid EXIF tag should simply be ignored or issue a warning message, instead of crash the whole program at once.
### Environment info
- `datasets` version: 3.6.0
- Platform: Linux-6.5.0-18-generic-x86_64-with-glibc2.35
- Python version: 3.12.11
- `huggingface_hub` version: 0.33.0
- PyArrow version: 20.0.0
- Pandas version: 2.3.0
- `fsspec` version: 2025.3.0 | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7668/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7668/timeline | null | null | null | null |
https://api.github.com/repos/huggingface/datasets/issues/7667 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7667/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7667/comments | https://api.github.com/repos/huggingface/datasets/issues/7667/events | https://github.com/huggingface/datasets/pull/7667 | 3,196,251,707 | PR_kwDODunzps6dGmm8 | 7,667 | Fix infer list of images | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7667). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-07-02T15:07:58Z | 2025-07-02T15:10:28Z | 2025-07-02T15:08:03Z | MEMBER | null | null | null | cc @kashif | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7667/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7667/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7667.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7667",
"merged_at": "2025-07-02T15:08:03Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7667.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7667"
} |
https://api.github.com/repos/huggingface/datasets/issues/7666 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7666/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7666/comments | https://api.github.com/repos/huggingface/datasets/issues/7666/events | https://github.com/huggingface/datasets/pull/7666 | 3,196,220,722 | PR_kwDODunzps6dGf7E | 7,666 | Backward compat list feature | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7666). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-07-02T14:58:00Z | 2025-07-02T15:00:37Z | 2025-07-02T14:59:40Z | MEMBER | null | null | null | cc @kashif | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 1,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7666/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7666/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7666.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7666",
"merged_at": "2025-07-02T14:59:40Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7666.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7666"
} |
https://api.github.com/repos/huggingface/datasets/issues/7665 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7665/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7665/comments | https://api.github.com/repos/huggingface/datasets/issues/7665/events | https://github.com/huggingface/datasets/issues/7665 | 3,193,239,955 | I_kwDODunzps6-VPmT | 7,665 | Function load_dataset() misinterprets string field content as part of dataset schema when dealing with `.jsonl` files | {
"avatar_url": "https://avatars.githubusercontent.com/u/1151198?v=4",
"events_url": "https://api.github.com/users/zdzichukowalski/events{/privacy}",
"followers_url": "https://api.github.com/users/zdzichukowalski/followers",
"following_url": "https://api.github.com/users/zdzichukowalski/following{/other_user}",
"gists_url": "https://api.github.com/users/zdzichukowalski/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/zdzichukowalski",
"id": 1151198,
"login": "zdzichukowalski",
"node_id": "MDQ6VXNlcjExNTExOTg=",
"organizations_url": "https://api.github.com/users/zdzichukowalski/orgs",
"received_events_url": "https://api.github.com/users/zdzichukowalski/received_events",
"repos_url": "https://api.github.com/users/zdzichukowalski/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/zdzichukowalski/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/zdzichukowalski/subscriptions",
"type": "User",
"url": "https://api.github.com/users/zdzichukowalski",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"Somehow I created the issue twice🙈 This one is an exact duplicate of #7664."
] | 2025-07-01T17:14:53Z | 2025-07-01T17:17:48Z | 2025-07-01T17:17:48Z | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
When loading a `.jsonl` file using `load_dataset("json", data_files="data.jsonl", split="train")`, the function misinterprets the content of a string field as if it were part of the dataset schema.
In my case there is a field `body:` with a string value
```
"### Describe the bug (...) ,action: string, datetime: timestamp[s], author: string, (...) Pandas version: 1.3.4"
```
As a result, I got an exception
```
"TypeError: Couldn't cast array of type timestamp[s] to null".
```
Full stack-trace in the attached file below.
I also attach a minimized dataset (data.json, a single entry) that reproduces the error.
**Observations**(on the minimal example):
- if I remove _all fields before_ `body`, a different error appears,
- if I remove _all fields after_ `body`, yet another error appears,
- if `body` is _the only field_, the error disappears.
So this might be one complex bug or several edge cases interacting. I haven’t dug deeper.
Also changing the file extension to `.json` or `.txt` avoids the problem. This suggests **a possible workaround** for the general case: convert `.jsonl` to `.json`. Though I haven’t verified correctness of that workaround yet.
Anyway my understanding is that `load_dataset` with first argument set to "json" should properly handle `.jsonl` files. Correct me if I'm wrong.
[stack_trace.txt](https://github.com/user-attachments/files/21004153/stack_trace.txt)
[data.json](https://github.com/user-attachments/files/21004164/data.json)
P.S.
I discovered this while going through the HuggingFace tutorial. Specifically [this part](https://huggingface.co/learn/llm-course/chapter5/5?fw=pt).I will try to inform the tutorial team about this issue, as it can be a showstopper for young 🤗 adepts.
### Steps to reproduce the bug
1. Download attached [data.json](https://github.com/user-attachments/files/21004164/data.json) file.
2. Run the following code which should work correctly:
```
from datasets import load_dataset
load_dataset("json", data_files="data.json", split="train")
```
3. Change extension of the `data` file to `.jsonl` and run:
```
from datasets import load_dataset
load_dataset("json", data_files="data.jsonl", split="train")
```
This will trigger an error like the one in the attached [stack_trace.txt](https://github.com/user-attachments/files/21004153/stack_trace.txt).
One can also try removing fields before the `body` field and after it. These actions give different errors.
### Expected behavior
Parsing data in `.jsonl` format should yield the same result as parsing the same data in `.json` format. In any case, the content of a string field should never be interpreted as part of the dataset schema.
### Environment info
datasets version: _3.6.0_
pyarrow version: _20.0.0_
Python version: _3.11.9_
platform version: _macOS-15.5-arm64-arm-64bit_ | {
"avatar_url": "https://avatars.githubusercontent.com/u/1151198?v=4",
"events_url": "https://api.github.com/users/zdzichukowalski/events{/privacy}",
"followers_url": "https://api.github.com/users/zdzichukowalski/followers",
"following_url": "https://api.github.com/users/zdzichukowalski/following{/other_user}",
"gists_url": "https://api.github.com/users/zdzichukowalski/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/zdzichukowalski",
"id": 1151198,
"login": "zdzichukowalski",
"node_id": "MDQ6VXNlcjExNTExOTg=",
"organizations_url": "https://api.github.com/users/zdzichukowalski/orgs",
"received_events_url": "https://api.github.com/users/zdzichukowalski/received_events",
"repos_url": "https://api.github.com/users/zdzichukowalski/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/zdzichukowalski/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/zdzichukowalski/subscriptions",
"type": "User",
"url": "https://api.github.com/users/zdzichukowalski",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7665/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7665/timeline | null | duplicate | null | null |
https://api.github.com/repos/huggingface/datasets/issues/7664 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7664/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7664/comments | https://api.github.com/repos/huggingface/datasets/issues/7664/events | https://github.com/huggingface/datasets/issues/7664 | 3,193,239,035 | I_kwDODunzps6-VPX7 | 7,664 | Function load_dataset() misinterprets string field content as part of dataset schema when dealing with `.jsonl` files | {
"avatar_url": "https://avatars.githubusercontent.com/u/1151198?v=4",
"events_url": "https://api.github.com/users/zdzichukowalski/events{/privacy}",
"followers_url": "https://api.github.com/users/zdzichukowalski/followers",
"following_url": "https://api.github.com/users/zdzichukowalski/following{/other_user}",
"gists_url": "https://api.github.com/users/zdzichukowalski/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/zdzichukowalski",
"id": 1151198,
"login": "zdzichukowalski",
"node_id": "MDQ6VXNlcjExNTExOTg=",
"organizations_url": "https://api.github.com/users/zdzichukowalski/orgs",
"received_events_url": "https://api.github.com/users/zdzichukowalski/received_events",
"repos_url": "https://api.github.com/users/zdzichukowalski/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/zdzichukowalski/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/zdzichukowalski/subscriptions",
"type": "User",
"url": "https://api.github.com/users/zdzichukowalski",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"Hey @zdzichukowalski, I was not able to reproduce this on python 3.11.9 and datasets 3.6.0. The contents of \"body\" are correctly parsed as a string and no other fields like timestamps are created. Could you try reproducing this in a fresh environment, or posting the complete code where you encountered that stack... | 2025-07-01T17:14:32Z | 2025-07-09T13:14:11Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
When loading a `.jsonl` file using `load_dataset("json", data_files="data.jsonl", split="train")`, the function misinterprets the content of a string field as if it were part of the dataset schema.
In my case there is a field `body:` with a string value
```
"### Describe the bug (...) ,action: string, datetime: timestamp[s], author: string, (...) Pandas version: 1.3.4"
```
As a result, I got an exception
```
"TypeError: Couldn't cast array of type timestamp[s] to null".
```
Full stack-trace in the attached file below.
I also attach a minimized dataset (data.json, a single entry) that reproduces the error.
**Observations**(on the minimal example):
- if I remove _all fields before_ `body`, a different error appears,
- if I remove _all fields after_ `body`, yet another error appears,
- if `body` is _the only field_, the error disappears.
So this might be one complex bug or several edge cases interacting. I haven’t dug deeper.
Also changing the file extension to `.json` or `.txt` avoids the problem. This suggests **a possible workaround** for the general case: convert `.jsonl` to `.json`. Though I haven’t verified correctness of that workaround yet.
Anyway my understanding is that `load_dataset` with first argument set to "json" should properly handle `.jsonl` files. Correct me if I'm wrong.
[stack_trace.txt](https://github.com/user-attachments/files/21004153/stack_trace.txt)
[data.json](https://github.com/user-attachments/files/21004164/data.json)
P.S.
I discovered this while going through the HuggingFace tutorial. Specifically [this part](https://huggingface.co/learn/llm-course/chapter5/5?fw=pt). I will try to inform the tutorial team about this issue, as it can be a showstopper for young 🤗 adepts.
### Steps to reproduce the bug
1. Download attached [data.json](https://github.com/user-attachments/files/21004164/data.json) file.
2. Run the following code which should work correctly:
```
from datasets import load_dataset
load_dataset("json", data_files="data.json", split="train")
```
3. Change extension of the `data` file to `.jsonl` and run:
```
from datasets import load_dataset
load_dataset("json", data_files="data.jsonl", split="train")
```
This will trigger an error like the one in the attached [stack_trace.txt](https://github.com/user-attachments/files/21004153/stack_trace.txt).
One can also try removing fields before the `body` field and after it. These actions give different errors.
### Expected behavior
Parsing data in `.jsonl` format should yield the same result as parsing the same data in `.json` format. In any case, the content of a string field should never be interpreted as part of the dataset schema.
### Environment info
datasets version: _3.6.0_
pyarrow version: _20.0.0_
Python version: _3.11.9_
platform version: _macOS-15.5-arm64-arm-64bit_ | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7664/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7664/timeline | null | null | null | null |
https://api.github.com/repos/huggingface/datasets/issues/7663 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7663/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7663/comments | https://api.github.com/repos/huggingface/datasets/issues/7663/events | https://github.com/huggingface/datasets/pull/7663 | 3,192,582,371 | PR_kwDODunzps6c6aJF | 7,663 | Custom metadata filenames | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7663). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2025-07-01T13:50:36Z | 2025-07-01T13:58:41Z | 2025-07-01T13:58:39Z | MEMBER | null | null | null | example: https://huggingface.co/datasets/lhoestq/overlapping-subsets-imagefolder/tree/main
To make multiple subsets for an imagefolder (one metadata file per subset), e.g.
```yaml
configs:
- config_name: default
metadata_filenames:
- metadata.csv
- config_name: other
metadata_filenames:
- metadata2.csv
``` | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 1,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7663/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7663/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7663.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7663",
"merged_at": "2025-07-01T13:58:39Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7663.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7663"
} |
https://api.github.com/repos/huggingface/datasets/issues/7662 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7662/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7662/comments | https://api.github.com/repos/huggingface/datasets/issues/7662/events | https://github.com/huggingface/datasets/issues/7662 | 3,190,805,531 | I_kwDODunzps6-L9Qb | 7,662 | Applying map after transform with multiprocessing will cause OOM | {
"avatar_url": "https://avatars.githubusercontent.com/u/26482910?v=4",
"events_url": "https://api.github.com/users/JunjieLl/events{/privacy}",
"followers_url": "https://api.github.com/users/JunjieLl/followers",
"following_url": "https://api.github.com/users/JunjieLl/following{/other_user}",
"gists_url": "https://api.github.com/users/JunjieLl/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/JunjieLl",
"id": 26482910,
"login": "JunjieLl",
"node_id": "MDQ6VXNlcjI2NDgyOTEw",
"organizations_url": "https://api.github.com/users/JunjieLl/orgs",
"received_events_url": "https://api.github.com/users/JunjieLl/received_events",
"repos_url": "https://api.github.com/users/JunjieLl/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/JunjieLl/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/JunjieLl/subscriptions",
"type": "User",
"url": "https://api.github.com/users/JunjieLl",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"Hi ! `add_column` loads the full column data in memory:\n\nhttps://github.com/huggingface/datasets/blob/bfa497b1666f4c58bd231c440d8b92f9859f3a58/src/datasets/arrow_dataset.py#L6021-L6021\n\na workaround to add the new column is to include the new data in the map() function instead, which only loads one batch at a ... | 2025-07-01T05:45:57Z | 2025-07-10T06:17:40Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
I have a 30TB dataset. When I perform add_column and cast_column operations on it and then execute a multiprocessing map, it results in an OOM (Out of Memory) error. However, if I skip the add_column and cast_column steps and directly run the map, there is no OOM. After debugging step by step, I found that the OOM is caused at this point, and I suspect it’s because the add_column and cast_column operations are not cached, which causes the entire dataset to be loaded in each subprocess, leading to the OOM. The critical line of code is: https://github.com/huggingface/datasets/blob/e71b0b19d79c7531f9b9bea7c09916b5f6157f42/src/datasets/utils/py_utils.py#L607
Note num_process=1 would not cause OOM. I'm confused.
### Steps to reproduce the bug
For reproduce, you can load dataset and set cache_dir (for caching): amphion/Emilia-Dataset which is a veru large datasets that RAM can not fits.
And apply the map with multiprocessing after a transform operation (e.g. add_column, cast_column).
As long as num_process>1, it must cause OOM.
### Expected behavior
It should not cause OOM.
### Environment info
- `datasets` version: 3.6.0
- Platform: Linux-5.10.134-16.101.al8.x86_64-x86_64-with-glibc2.35
- Python version: 3.10.12
- `huggingface_hub` version: 0.33.1
- PyArrow version: 20.0.0
- Pandas version: 2.3.0
- `fsspec` version: 2024.6.1 | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7662/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7662/timeline | null | null | null | null |
https://api.github.com/repos/huggingface/datasets/issues/7661 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7661/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7661/comments | https://api.github.com/repos/huggingface/datasets/issues/7661/events | https://github.com/huggingface/datasets/pull/7661 | 3,190,408,237 | PR_kwDODunzps6czBDi | 7,661 | fix del tqdm lock error | {
"avatar_url": "https://avatars.githubusercontent.com/u/44766273?v=4",
"events_url": "https://api.github.com/users/Hypothesis-Z/events{/privacy}",
"followers_url": "https://api.github.com/users/Hypothesis-Z/followers",
"following_url": "https://api.github.com/users/Hypothesis-Z/following{/other_user}",
"gists_url": "https://api.github.com/users/Hypothesis-Z/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/Hypothesis-Z",
"id": 44766273,
"login": "Hypothesis-Z",
"node_id": "MDQ6VXNlcjQ0NzY2Mjcz",
"organizations_url": "https://api.github.com/users/Hypothesis-Z/orgs",
"received_events_url": "https://api.github.com/users/Hypothesis-Z/received_events",
"repos_url": "https://api.github.com/users/Hypothesis-Z/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/Hypothesis-Z/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Hypothesis-Z/subscriptions",
"type": "User",
"url": "https://api.github.com/users/Hypothesis-Z",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [] | 2025-07-01T02:04:02Z | 2025-07-08T01:38:46Z | null | NONE | null | null | null | fixes https://github.com/huggingface/datasets/issues/7660 | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7661/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7661/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7661.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7661",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/7661.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7661"
} |
https://api.github.com/repos/huggingface/datasets/issues/7660 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7660/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7660/comments | https://api.github.com/repos/huggingface/datasets/issues/7660/events | https://github.com/huggingface/datasets/issues/7660 | 3,189,028,251 | I_kwDODunzps6-FLWb | 7,660 | AttributeError: type object 'tqdm' has no attribute '_lock' | {
"avatar_url": "https://avatars.githubusercontent.com/u/44766273?v=4",
"events_url": "https://api.github.com/users/Hypothesis-Z/events{/privacy}",
"followers_url": "https://api.github.com/users/Hypothesis-Z/followers",
"following_url": "https://api.github.com/users/Hypothesis-Z/following{/other_user}",
"gists_url": "https://api.github.com/users/Hypothesis-Z/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/Hypothesis-Z",
"id": 44766273,
"login": "Hypothesis-Z",
"node_id": "MDQ6VXNlcjQ0NzY2Mjcz",
"organizations_url": "https://api.github.com/users/Hypothesis-Z/orgs",
"received_events_url": "https://api.github.com/users/Hypothesis-Z/received_events",
"repos_url": "https://api.github.com/users/Hypothesis-Z/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/Hypothesis-Z/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Hypothesis-Z/subscriptions",
"type": "User",
"url": "https://api.github.com/users/Hypothesis-Z",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [
"Deleting a class (**not instance**) attribute might be invalid in this case, which is `tqdm` doing in `ensure_lock`.\n\n```python\nfrom tqdm import tqdm as old_tqdm\n\nclass tqdm1(old_tqdm):\n def __delattr__(self, attr):\n try:\n super().__delattr__(attr)\n except AttributeError:\n ... | 2025-06-30T15:57:16Z | 2025-07-03T15:14:27Z | null | NONE | null | null | {
"completed": 0,
"percent_completed": 0,
"total": 0
} | ### Describe the bug
`AttributeError: type object 'tqdm' has no attribute '_lock'`
It occurs when I'm trying to load datasets in thread pool.
Issue https://github.com/huggingface/datasets/issues/6066 and PR https://github.com/huggingface/datasets/pull/6067 https://github.com/huggingface/datasets/pull/6068 tried to fix this.
### Steps to reproduce the bug
Will have to try several times to reproduce the error because it is concerned with threads.
1. save some datasets for test
```pythonfrom datasets import Dataset, DatasetDict
import os
os.makedirs("test_dataset_shards", exist_ok=True)
for i in range(10):
data = Dataset.from_dict({"text": [f"example {j}" for j in range(1000000)]})
data = DatasetDict({'train': data})
data.save_to_disk(f"test_dataset_shards/shard_{i}")
```
2. load them in a thread pool
```python
from datasets import load_from_disk
from tqdm import tqdm
from concurrent.futures import ThreadPoolExecutor, as_completed
import glob
datas = glob.glob('test_dataset_shards/shard_*')
with ThreadPoolExecutor(max_workers=10) as pool:
futures = [pool.submit(load_from_disk, it) for it in datas]
datas = []
for future in tqdm(as_completed(futures), total=len(futures)):
datas.append(future.result())
```
### Expected behavior
no exception raised
### Environment info
datasets==2.19.0
python==3.10 | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7660/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7660/timeline | null | null | null | null |
https://api.github.com/repos/huggingface/datasets/issues/7659 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7659/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7659/comments | https://api.github.com/repos/huggingface/datasets/issues/7659/events | https://github.com/huggingface/datasets/pull/7659 | 3,187,882,217 | PR_kwDODunzps6cqkou | 7,659 | Update the beans dataset link in Preprocess | {
"avatar_url": "https://avatars.githubusercontent.com/u/5434867?v=4",
"events_url": "https://api.github.com/users/HJassar/events{/privacy}",
"followers_url": "https://api.github.com/users/HJassar/followers",
"following_url": "https://api.github.com/users/HJassar/following{/other_user}",
"gists_url": "https://api.github.com/users/HJassar/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/HJassar",
"id": 5434867,
"login": "HJassar",
"node_id": "MDQ6VXNlcjU0MzQ4Njc=",
"organizations_url": "https://api.github.com/users/HJassar/orgs",
"received_events_url": "https://api.github.com/users/HJassar/received_events",
"repos_url": "https://api.github.com/users/HJassar/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/HJassar/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/HJassar/subscriptions",
"type": "User",
"url": "https://api.github.com/users/HJassar",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [] | 2025-06-30T09:58:44Z | 2025-07-07T08:38:19Z | 2025-07-01T14:01:42Z | CONTRIBUTOR | null | null | null | In the Preprocess tutorial, the to "the beans dataset" is incorrect. Fixed. | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/lhoestq",
"id": 42851186,
"login": "lhoestq",
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"organizations_url": "https://api.github.com/users/lhoestq/orgs",
"received_events_url": "https://api.github.com/users/lhoestq/received_events",
"repos_url": "https://api.github.com/users/lhoestq/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions",
"type": "User",
"url": "https://api.github.com/users/lhoestq",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7659/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7659/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7659.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7659",
"merged_at": "2025-07-01T14:01:42Z",
"patch_url": "https://github.com/huggingface/datasets/pull/7659.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7659"
} |
https://api.github.com/repos/huggingface/datasets/issues/7658 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7658/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7658/comments | https://api.github.com/repos/huggingface/datasets/issues/7658/events | https://github.com/huggingface/datasets/pull/7658 | 3,187,800,504 | PR_kwDODunzps6cqTMs | 7,658 | Fix: Prevent loss of info.features and column_names in IterableDatasetDict.map when features is None | {
"avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4",
"events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}",
"followers_url": "https://api.github.com/users/ArjunJagdale/followers",
"following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}",
"gists_url": "https://api.github.com/users/ArjunJagdale/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/ArjunJagdale",
"id": 142811259,
"login": "ArjunJagdale",
"node_id": "U_kgDOCIMgew",
"organizations_url": "https://api.github.com/users/ArjunJagdale/orgs",
"received_events_url": "https://api.github.com/users/ArjunJagdale/received_events",
"repos_url": "https://api.github.com/users/ArjunJagdale/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/ArjunJagdale/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ArjunJagdale/subscriptions",
"type": "User",
"url": "https://api.github.com/users/ArjunJagdale",
"user_view_type": "public"
} | [] | closed | false | null | [] | null | [
"Hi!\r\nI haven’t included a test for this change, as the fix is quite small and targeted.\r\nPlease let me know if you’d like a test for this case or if you’d prefer to handle it during review.\r\nThanks!",
"we can't know in advance the `features` after map() (it transforms the data !), so you can reuse the `fea... | 2025-06-30T09:31:12Z | 2025-07-01T16:26:30Z | 2025-07-01T16:26:12Z | CONTRIBUTOR | null | null | null | This PR fixes a bug where calling `IterableDatasetDict.map()` or `IterableDataset.map()` with the default `features=None` argument would overwrite the existing `info.features` attribute with `None`. This, in turn, caused the resulting dataset to lose its schema, breaking downstream usage of attributes like `column_names`.
Why
Previously, the code would always set `info.features = features`, even if `features` was `None`. When mapping with removal of columns or other transformations, this led to the destruction of the schema and caused failures in code that relied on the dataset schema being present.
How
We now update `info.features` only if `features` is not `None`. This preserves the original schema unless the user explicitly provides a new one.
Reference
Fixes #7568 | {
"avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4",
"events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}",
"followers_url": "https://api.github.com/users/ArjunJagdale/followers",
"following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}",
"gists_url": "https://api.github.com/users/ArjunJagdale/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/ArjunJagdale",
"id": 142811259,
"login": "ArjunJagdale",
"node_id": "U_kgDOCIMgew",
"organizations_url": "https://api.github.com/users/ArjunJagdale/orgs",
"received_events_url": "https://api.github.com/users/ArjunJagdale/received_events",
"repos_url": "https://api.github.com/users/ArjunJagdale/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/ArjunJagdale/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ArjunJagdale/subscriptions",
"type": "User",
"url": "https://api.github.com/users/ArjunJagdale",
"user_view_type": "public"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7658/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7658/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7658.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7658",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/7658.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7658"
} |
https://api.github.com/repos/huggingface/datasets/issues/7657 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7657/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7657/comments | https://api.github.com/repos/huggingface/datasets/issues/7657/events | https://github.com/huggingface/datasets/pull/7657 | 3,186,036,016 | PR_kwDODunzps6cks2E | 7,657 | feat: add subset_name as alias for name in load_dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4",
"events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}",
"followers_url": "https://api.github.com/users/ArjunJagdale/followers",
"following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}",
"gists_url": "https://api.github.com/users/ArjunJagdale/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/ArjunJagdale",
"id": 142811259,
"login": "ArjunJagdale",
"node_id": "U_kgDOCIMgew",
"organizations_url": "https://api.github.com/users/ArjunJagdale/orgs",
"received_events_url": "https://api.github.com/users/ArjunJagdale/received_events",
"repos_url": "https://api.github.com/users/ArjunJagdale/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/ArjunJagdale/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ArjunJagdale/subscriptions",
"type": "User",
"url": "https://api.github.com/users/ArjunJagdale",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [] | 2025-06-29T10:39:00Z | 2025-06-29T10:55:11Z | null | CONTRIBUTOR | null | null | null | fixes #7637
This PR introduces subset_name as a user-facing alias for the name (previously `config_name`) argument in load_dataset. It aligns terminology with the Hugging Face Hub UI (which shows “Subset”), reducing confusion for new users.
Supports `subset_name` in `load_dataset()`
Adds `.subset_name` property to DatasetBuilder
Maintains full backward compatibility
Raises clear error if name and `subset_name` conflict | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7657/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7657/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7657.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7657",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/7657.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7657"
} |
https://api.github.com/repos/huggingface/datasets/issues/7656 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7656/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7656/comments | https://api.github.com/repos/huggingface/datasets/issues/7656/events | https://github.com/huggingface/datasets/pull/7656 | 3,185,865,686 | PR_kwDODunzps6ckPHc | 7,656 | fix(iterable): ensure MappedExamplesIterable supports state_dict for resume | {
"avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4",
"events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}",
"followers_url": "https://api.github.com/users/ArjunJagdale/followers",
"following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}",
"gists_url": "https://api.github.com/users/ArjunJagdale/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/ArjunJagdale",
"id": 142811259,
"login": "ArjunJagdale",
"node_id": "U_kgDOCIMgew",
"organizations_url": "https://api.github.com/users/ArjunJagdale/orgs",
"received_events_url": "https://api.github.com/users/ArjunJagdale/received_events",
"repos_url": "https://api.github.com/users/ArjunJagdale/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/ArjunJagdale/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ArjunJagdale/subscriptions",
"type": "User",
"url": "https://api.github.com/users/ArjunJagdale",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [] | 2025-06-29T07:50:13Z | 2025-06-29T07:50:13Z | null | CONTRIBUTOR | null | null | null | Fixes #7630
### Problem
When calling `.map()` on an `IterableDataset`, resuming from a checkpoint skips a large number of samples. This is because `MappedExamplesIterable` did not implement `state_dict()` or `load_state_dict()`, so checkpointing was not properly delegated to the underlying iterable.
### What This PR Does
This patch adds:
```python
def state_dict(self):
return self.ex_iterable.state_dict()
def load_state_dict(self, state):
self.ex_iterable.load_state_dict(state)
```
to MappedExamplesIterable, so the wrapped base iterable's state can be saved and restored as expected.
Result
Using .map() no longer causes sample skipping after checkpoint resume.
Let me know if a dedicated test case is required — happy to add one! | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7656/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7656/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7656.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7656",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/7656.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7656"
} |
https://api.github.com/repos/huggingface/datasets/issues/7655 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7655/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7655/comments | https://api.github.com/repos/huggingface/datasets/issues/7655/events | https://github.com/huggingface/datasets/pull/7655 | 3,185,382,105 | PR_kwDODunzps6ci9oi | 7,655 | Added specific use cases in Improve Performace | {
"avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4",
"events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}",
"followers_url": "https://api.github.com/users/ArjunJagdale/followers",
"following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}",
"gists_url": "https://api.github.com/users/ArjunJagdale/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/ArjunJagdale",
"id": 142811259,
"login": "ArjunJagdale",
"node_id": "U_kgDOCIMgew",
"organizations_url": "https://api.github.com/users/ArjunJagdale/orgs",
"received_events_url": "https://api.github.com/users/ArjunJagdale/received_events",
"repos_url": "https://api.github.com/users/ArjunJagdale/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/ArjunJagdale/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ArjunJagdale/subscriptions",
"type": "User",
"url": "https://api.github.com/users/ArjunJagdale",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [] | 2025-06-28T19:00:32Z | 2025-06-28T19:00:32Z | null | CONTRIBUTOR | null | null | null | Fixes #2494 | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7655/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7655/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7655.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7655",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/7655.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7655"
} |
https://api.github.com/repos/huggingface/datasets/issues/7654 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7654/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7654/comments | https://api.github.com/repos/huggingface/datasets/issues/7654/events | https://github.com/huggingface/datasets/pull/7654 | 3,184,770,992 | PR_kwDODunzps6chPmz | 7,654 | fix(load): strip deprecated use_auth_token from config_kwargs | {
"avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4",
"events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}",
"followers_url": "https://api.github.com/users/ArjunJagdale/followers",
"following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}",
"gists_url": "https://api.github.com/users/ArjunJagdale/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/ArjunJagdale",
"id": 142811259,
"login": "ArjunJagdale",
"node_id": "U_kgDOCIMgew",
"organizations_url": "https://api.github.com/users/ArjunJagdale/orgs",
"received_events_url": "https://api.github.com/users/ArjunJagdale/received_events",
"repos_url": "https://api.github.com/users/ArjunJagdale/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/ArjunJagdale/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ArjunJagdale/subscriptions",
"type": "User",
"url": "https://api.github.com/users/ArjunJagdale",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [] | 2025-06-28T09:20:21Z | 2025-06-28T09:20:21Z | null | CONTRIBUTOR | null | null | null | Fixes #7504
This PR resolves a compatibility error when loading datasets via `load_dataset()` using outdated arguments like `use_auth_token`.
**What was happening:**
Users passing `use_auth_token` in `load_dataset(..., use_auth_token=...)` encountered a `ValueError`: BuilderConfig ParquetConfig(...) doesn't have a 'use_auth_token' key.
**Why:**
`use_auth_token` has been deprecated and removed from config definitions (replaced by `token`), but the `load_dataset()` function still forwarded it via `**config_kwargs` to BuilderConfigs, leading to unrecognized key errors.
**Fix:**
We now intercept and strip `use_auth_token` from `config_kwargs` inside `load_dataset`, replacing it with a warning:
```python
if "use_auth_token" in config_kwargs:
logger.warning("The 'use_auth_token' argument is deprecated. Please use 'token' instead.")
config_kwargs.pop("use_auth_token")
```
This ensures legacy compatibility while guiding users to switch to the token argument.
Let me know if you'd prefer a deprecation error instead of a warning. Thanks! | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7654/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7654/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7654.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7654",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/7654.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7654"
} |
https://api.github.com/repos/huggingface/datasets/issues/7653 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7653/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7653/comments | https://api.github.com/repos/huggingface/datasets/issues/7653/events | https://github.com/huggingface/datasets/pull/7653 | 3,184,746,093 | PR_kwDODunzps6chLmb | 7,653 | feat(load): fallback to `load_from_disk()` when loading a saved dataset directory | {
"avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4",
"events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}",
"followers_url": "https://api.github.com/users/ArjunJagdale/followers",
"following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}",
"gists_url": "https://api.github.com/users/ArjunJagdale/gists{/gist_id}",
"gravatar_id": "",
"html_url": "https://github.com/ArjunJagdale",
"id": 142811259,
"login": "ArjunJagdale",
"node_id": "U_kgDOCIMgew",
"organizations_url": "https://api.github.com/users/ArjunJagdale/orgs",
"received_events_url": "https://api.github.com/users/ArjunJagdale/received_events",
"repos_url": "https://api.github.com/users/ArjunJagdale/repos",
"site_admin": false,
"starred_url": "https://api.github.com/users/ArjunJagdale/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ArjunJagdale/subscriptions",
"type": "User",
"url": "https://api.github.com/users/ArjunJagdale",
"user_view_type": "public"
} | [] | open | false | null | [] | null | [] | 2025-06-28T08:47:36Z | 2025-06-28T08:47:36Z | null | CONTRIBUTOR | null | null | null | ### Related Issue
Fixes #7503
Partially addresses #5044 by allowing `load_dataset()` to auto-detect and gracefully delegate to `load_from_disk()` for locally saved datasets.
---
### What does this PR do?
This PR introduces a minimal fallback mechanism in `load_dataset()` that detects when the provided `path` points to a dataset saved using `save_to_disk()`, and automatically redirects to `load_from_disk()`.
#### 🐛 Before (unexpected metadata-only rows):
```python
ds = load_dataset("/path/to/saved_dataset")
# → returns rows with only internal metadata (_data_files, _fingerprint, etc.)
````
#### ✅ After (graceful fallback):
```python
ds = load_dataset("/path/to/saved_dataset")
# → logs a warning and internally switches to load_from_disk()
```
---
### Why is this useful?
* Prevents confusion when reloading local datasets saved via `save_to_disk()`.
* Enables smoother compatibility with frameworks (e.g., TRL, `lighteval`) that rely on `load_dataset()` calls.
* Fully backward-compatible — hub-based loading, custom builders, and streaming remain untouched.
| null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/7653/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/7653/timeline | null | null | false | {
"diff_url": "https://github.com/huggingface/datasets/pull/7653.diff",
"html_url": "https://github.com/huggingface/datasets/pull/7653",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/7653.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7653"
} |
End of preview. Expand in Data Studio
- Downloads last month
- 14