|
|
|
|
|
""" |
|
|
Create viktoroo/longbench-pro-128k-plus from caskcsg/LongBench-Pro by: |
|
|
- filtering to token_length in {"128k", "256k"} |
|
|
- keeping only fields: id, context |
|
|
- renaming context -> text |
|
|
- pushing the filtered dataset to the (already-existing) public repo |
|
|
- uploading this script and a hardcoded README.md into the same dataset repo |
|
|
|
|
|
Requirements: |
|
|
pip install -U datasets huggingface_hub |
|
|
|
|
|
Auth: |
|
|
export HF_TOKEN=... (must have write access to viktoroo/longbench-pro-128k-plus) |
|
|
""" |
|
|
|
|
|
from __future__ import annotations |
|
|
|
|
|
import os |
|
|
import sys |
|
|
import tempfile |
|
|
from pathlib import Path |
|
|
|
|
|
from datasets import load_dataset, DatasetDict |
|
|
from huggingface_hub import HfApi |
|
|
|
|
|
from dotenv import load_dotenv |
|
|
load_dotenv() |
|
|
|
|
|
|
|
|
SOURCE_DATASET = "caskcsg/LongBench-Pro" |
|
|
TARGET_REPO = "viktoroo/longbench-pro-128k-plus" |
|
|
ALLOWED_TOKEN_LENGTH = {"128k", "256k"} |
|
|
|
|
|
README_MD = """--- |
|
|
license: other |
|
|
language: |
|
|
- en |
|
|
- zh |
|
|
tags: |
|
|
- long-context |
|
|
- benchmark |
|
|
- evaluation |
|
|
- rag |
|
|
pretty_name: LongBench Pro 128k+ |
|
|
--- |
|
|
|
|
|
# LongBench Pro 128k+ |
|
|
|
|
|
This dataset is a filtered subset of **LongBench Pro** (`caskcsg/LongBench-Pro`). |
|
|
|
|
|
## What is included |
|
|
|
|
|
Only examples whose `token_length` field is one of: |
|
|
|
|
|
- `128k` |
|
|
- `256k` |
|
|
|
|
|
## Columns |
|
|
|
|
|
This repo keeps only: |
|
|
|
|
|
- `id`: example identifier (copied from source) |
|
|
- `text`: the original `context` field (renamed from `context` → `text`) |
|
|
|
|
|
All other fields from the source dataset are dropped. |
|
|
|
|
|
## Intended use |
|
|
|
|
|
Use this dataset when you want to benchmark long-context behavior specifically at **≥128k** length buckets, while keeping the input surface minimal (`id`, `text`). |
|
|
|
|
|
## Provenance / attribution |
|
|
|
|
|
Source dataset: `caskcsg/LongBench-Pro`. |
|
|
|
|
|
This repo contains a derived subset. Please consult the source dataset card for: |
|
|
- full task definitions |
|
|
- original annotations/fields |
|
|
- licensing/usage terms |
|
|
|
|
|
## Reproducibility |
|
|
|
|
|
The filtering logic and transformation used to build this dataset are contained in `create_dataset.py` in this repo. |
|
|
""" |
|
|
|
|
|
|
|
|
def require_token() -> str: |
|
|
token = os.environ.get("HF_TOKEN") or os.environ.get("HUGGINGFACE_TOKEN") |
|
|
if not token: |
|
|
raise RuntimeError("Missing HF_TOKEN (or HUGGINGFACE_TOKEN) env var.") |
|
|
return token |
|
|
|
|
|
|
|
|
def filter_and_project(ds: DatasetDict) -> DatasetDict: |
|
|
out = DatasetDict() |
|
|
for split, d in ds.items(): |
|
|
if "token_length" not in d.column_names: |
|
|
raise RuntimeError(f"Split '{split}' has no 'token_length' column.") |
|
|
|
|
|
if "context" not in d.column_names: |
|
|
raise RuntimeError(f"Split '{split}' has no 'context' column.") |
|
|
|
|
|
if "id" not in d.column_names: |
|
|
raise RuntimeError(f"Split '{split}' has no 'id' column.") |
|
|
|
|
|
d2 = d.filter(lambda ex: ex["token_length"] in ALLOWED_TOKEN_LENGTH) |
|
|
|
|
|
|
|
|
d2 = d2.select_columns(["id", "context"]).rename_column("context", "text") |
|
|
out[split] = d2 |
|
|
|
|
|
return out |
|
|
|
|
|
|
|
|
def main() -> int: |
|
|
token = require_token() |
|
|
|
|
|
print(f"Loading source dataset: {SOURCE_DATASET}") |
|
|
ds = load_dataset(SOURCE_DATASET) |
|
|
|
|
|
print("Filtering and projecting columns...") |
|
|
out = filter_and_project(ds) |
|
|
|
|
|
|
|
|
for split, d in out.items(): |
|
|
print(f"Split '{split}': {len(d)} rows; columns={d.column_names}") |
|
|
|
|
|
print(f"Pushing dataset to hub: {TARGET_REPO}") |
|
|
out.push_to_hub( |
|
|
TARGET_REPO, |
|
|
token=token, |
|
|
private=False, |
|
|
commit_message="Create/update filtered LongBench Pro subset (128k, 256k) with id+text", |
|
|
) |
|
|
|
|
|
|
|
|
api = HfApi(token=token) |
|
|
|
|
|
|
|
|
script_path = Path(__file__).resolve() |
|
|
|
|
|
with tempfile.TemporaryDirectory() as td: |
|
|
td_path = Path(td) |
|
|
readme_path = td_path / "README.md" |
|
|
readme_path.write_text(README_MD, encoding="utf-8") |
|
|
|
|
|
print("Uploading README.md...") |
|
|
api.upload_file( |
|
|
path_or_fileobj=str(readme_path), |
|
|
path_in_repo="README.md", |
|
|
repo_id=TARGET_REPO, |
|
|
repo_type="dataset", |
|
|
commit_message="Add dataset README", |
|
|
) |
|
|
|
|
|
print("Uploading create_dataset.py...") |
|
|
api.upload_file( |
|
|
path_or_fileobj=str(script_path), |
|
|
path_in_repo="create_dataset.py", |
|
|
repo_id=TARGET_REPO, |
|
|
repo_type="dataset", |
|
|
commit_message="Add dataset creation script", |
|
|
) |
|
|
|
|
|
print("Done.") |
|
|
return 0 |
|
|
|
|
|
|
|
|
if __name__ == "__main__": |
|
|
try: |
|
|
raise SystemExit(main()) |
|
|
except Exception as e: |
|
|
print(f"ERROR: {e}", file=sys.stderr) |
|
|
raise |
|
|
|