|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
"""Contains commands to interact with datasets on the Hugging Face Hub. |
|
|
|
|
|
Usage: |
|
|
# list datasets on the Hub |
|
|
hf datasets ls |
|
|
|
|
|
# list datasets with a search query |
|
|
hf datasets ls --search "code" |
|
|
|
|
|
# get info about a dataset |
|
|
hf datasets info HuggingFaceFW/fineweb |
|
|
""" |
|
|
|
|
|
import enum |
|
|
import json |
|
|
from typing import Annotated, Optional, get_args |
|
|
|
|
|
import typer |
|
|
|
|
|
from huggingface_hub.errors import RepositoryNotFoundError, RevisionNotFoundError |
|
|
from huggingface_hub.hf_api import DatasetSort_T, ExpandDatasetProperty_T |
|
|
from huggingface_hub.utils import ANSI |
|
|
|
|
|
from ._cli_utils import ( |
|
|
AuthorOpt, |
|
|
FilterOpt, |
|
|
LimitOpt, |
|
|
RevisionOpt, |
|
|
SearchOpt, |
|
|
TokenOpt, |
|
|
get_hf_api, |
|
|
make_expand_properties_parser, |
|
|
repo_info_to_dict, |
|
|
typer_factory, |
|
|
) |
|
|
|
|
|
|
|
|
_EXPAND_PROPERTIES = sorted(get_args(ExpandDatasetProperty_T)) |
|
|
_SORT_OPTIONS = get_args(DatasetSort_T) |
|
|
DatasetSortEnum = enum.Enum("DatasetSortEnum", {s: s for s in _SORT_OPTIONS}, type=str) |
|
|
|
|
|
|
|
|
ExpandOpt = Annotated[ |
|
|
Optional[str], |
|
|
typer.Option( |
|
|
help=f"Comma-separated properties to expand. Example: '--expand=downloads,likes,tags'. Valid: {', '.join(_EXPAND_PROPERTIES)}.", |
|
|
callback=make_expand_properties_parser(_EXPAND_PROPERTIES), |
|
|
), |
|
|
] |
|
|
|
|
|
|
|
|
datasets_cli = typer_factory(help="Interact with datasets on the Hub.") |
|
|
|
|
|
|
|
|
@datasets_cli.command("ls") |
|
|
def datasets_ls( |
|
|
search: SearchOpt = None, |
|
|
author: AuthorOpt = None, |
|
|
filter: FilterOpt = None, |
|
|
sort: Annotated[ |
|
|
Optional[DatasetSortEnum], |
|
|
typer.Option(help="Sort results."), |
|
|
] = None, |
|
|
limit: LimitOpt = 10, |
|
|
expand: ExpandOpt = None, |
|
|
token: TokenOpt = None, |
|
|
) -> None: |
|
|
"""List datasets on the Hub.""" |
|
|
api = get_hf_api(token=token) |
|
|
sort_key = sort.value if sort else None |
|
|
results = [ |
|
|
repo_info_to_dict(dataset_info) |
|
|
for dataset_info in api.list_datasets( |
|
|
filter=filter, author=author, search=search, sort=sort_key, limit=limit, expand=expand |
|
|
) |
|
|
] |
|
|
print(json.dumps(results, indent=2)) |
|
|
|
|
|
|
|
|
@datasets_cli.command("info") |
|
|
def datasets_info( |
|
|
dataset_id: Annotated[str, typer.Argument(help="The dataset ID (e.g. `username/repo-name`).")], |
|
|
revision: RevisionOpt = None, |
|
|
expand: ExpandOpt = None, |
|
|
token: TokenOpt = None, |
|
|
) -> None: |
|
|
"""Get info about a dataset on the Hub.""" |
|
|
api = get_hf_api(token=token) |
|
|
try: |
|
|
info = api.dataset_info(repo_id=dataset_id, revision=revision, expand=expand) |
|
|
except RepositoryNotFoundError: |
|
|
print(f"Dataset {ANSI.bold(dataset_id)} not found.") |
|
|
raise typer.Exit(code=1) |
|
|
except RevisionNotFoundError: |
|
|
print(f"Revision {ANSI.bold(str(revision))} not found on {ANSI.bold(dataset_id)}.") |
|
|
raise typer.Exit(code=1) |
|
|
print(json.dumps(repo_info_to_dict(info), indent=2)) |
|
|
|