| | """ |
| | Implementation of a custom transfer agent for the transfer type "multipart" for |
| | git-lfs. |
| | |
| | Inspired by: |
| | github.com/cbartz/git-lfs-swift-transfer-agent/blob/master/git_lfs_swift_transfer.py |
| | |
| | Spec is: github.com/git-lfs/git-lfs/blob/master/docs/custom-transfers.md |
| | |
| | |
| | To launch debugger while developing: |
| | |
| | ``` [lfs "customtransfer.multipart"] |
| | path = /path/to/huggingface_hub/.env/bin/python args = -m debugpy --listen 5678 |
| | --wait-for-client |
| | /path/to/huggingface_hub/src/huggingface_hub/commands/huggingface_cli.py |
| | lfs-multipart-upload ```""" |
| |
|
| | import json |
| | import os |
| | import subprocess |
| | import sys |
| | from argparse import _SubParsersAction |
| | from typing import Dict, List, Optional |
| |
|
| | from huggingface_hub.commands import BaseHuggingfaceCLICommand |
| | from huggingface_hub.lfs import LFS_MULTIPART_UPLOAD_COMMAND |
| |
|
| | from ..utils import get_session, hf_raise_for_status, logging |
| | from ..utils._lfs import SliceFileObj |
| |
|
| |
|
| | logger = logging.get_logger(__name__) |
| |
|
| |
|
| | class LfsCommands(BaseHuggingfaceCLICommand): |
| | """ |
| | Implementation of a custom transfer agent for the transfer type "multipart" |
| | for git-lfs. This lets users upload large files >5GB 🔥. Spec for LFS custom |
| | transfer agent is: |
| | https://github.com/git-lfs/git-lfs/blob/master/docs/custom-transfers.md |
| | |
| | This introduces two commands to the CLI: |
| | |
| | 1. $ hf lfs-enable-largefiles |
| | |
| | This should be executed once for each model repo that contains a model file |
| | >5GB. It's documented in the error message you get if you just try to git |
| | push a 5GB file without having enabled it before. |
| | |
| | 2. $ hf lfs-multipart-upload |
| | |
| | This command is called by lfs directly and is not meant to be called by the |
| | user. |
| | """ |
| |
|
| | @staticmethod |
| | def register_subcommand(parser: _SubParsersAction): |
| | enable_parser = parser.add_parser("lfs-enable-largefiles", add_help=False) |
| | enable_parser.add_argument("path", type=str, help="Local path to repository you want to configure.") |
| | enable_parser.set_defaults(func=lambda args: LfsEnableCommand(args)) |
| |
|
| | |
| | upload_parser = parser.add_parser(LFS_MULTIPART_UPLOAD_COMMAND, add_help=False) |
| | upload_parser.set_defaults(func=lambda args: LfsUploadCommand(args)) |
| |
|
| |
|
| | class LfsEnableCommand: |
| | def __init__(self, args): |
| | self.args = args |
| |
|
| | def run(self): |
| | local_path = os.path.abspath(self.args.path) |
| | if not os.path.isdir(local_path): |
| | print("This does not look like a valid git repo.") |
| | exit(1) |
| | subprocess.run( |
| | "git config lfs.customtransfer.multipart.path hf".split(), |
| | check=True, |
| | cwd=local_path, |
| | ) |
| | subprocess.run( |
| | f"git config lfs.customtransfer.multipart.args {LFS_MULTIPART_UPLOAD_COMMAND}".split(), |
| | check=True, |
| | cwd=local_path, |
| | ) |
| | print("Local repo set up for largefiles") |
| |
|
| |
|
| | def write_msg(msg: Dict): |
| | """Write out the message in Line delimited JSON.""" |
| | msg_str = json.dumps(msg) + "\n" |
| | sys.stdout.write(msg_str) |
| | sys.stdout.flush() |
| |
|
| |
|
| | def read_msg() -> Optional[Dict]: |
| | """Read Line delimited JSON from stdin.""" |
| | msg = json.loads(sys.stdin.readline().strip()) |
| |
|
| | if "terminate" in (msg.get("type"), msg.get("event")): |
| | |
| | return None |
| |
|
| | if msg.get("event") not in ("download", "upload"): |
| | logger.critical("Received unexpected message") |
| | sys.exit(1) |
| |
|
| | return msg |
| |
|
| |
|
| | class LfsUploadCommand: |
| | def __init__(self, args) -> None: |
| | self.args = args |
| |
|
| | def run(self) -> None: |
| | |
| | |
| | |
| | init_msg = json.loads(sys.stdin.readline().strip()) |
| | if not (init_msg.get("event") == "init" and init_msg.get("operation") == "upload"): |
| | write_msg({"error": {"code": 32, "message": "Wrong lfs init operation"}}) |
| | sys.exit(1) |
| |
|
| | |
| | |
| | |
| | |
| | write_msg({}) |
| |
|
| | |
| | |
| | while True: |
| | msg = read_msg() |
| | if msg is None: |
| | |
| | |
| | |
| | |
| | sys.exit(0) |
| |
|
| | oid = msg["oid"] |
| | filepath = msg["path"] |
| | completion_url = msg["action"]["href"] |
| | header = msg["action"]["header"] |
| | chunk_size = int(header.pop("chunk_size")) |
| | presigned_urls: List[str] = list(header.values()) |
| |
|
| | |
| | |
| | |
| | write_msg( |
| | { |
| | "event": "progress", |
| | "oid": oid, |
| | "bytesSoFar": 1, |
| | "bytesSinceLast": 0, |
| | } |
| | ) |
| |
|
| | parts = [] |
| | with open(filepath, "rb") as file: |
| | for i, presigned_url in enumerate(presigned_urls): |
| | with SliceFileObj( |
| | file, |
| | seek_from=i * chunk_size, |
| | read_limit=chunk_size, |
| | ) as data: |
| | r = get_session().put(presigned_url, data=data) |
| | hf_raise_for_status(r) |
| | parts.append( |
| | { |
| | "etag": r.headers.get("etag"), |
| | "partNumber": i + 1, |
| | } |
| | ) |
| | |
| | |
| | write_msg( |
| | { |
| | "event": "progress", |
| | "oid": oid, |
| | "bytesSoFar": (i + 1) * chunk_size, |
| | "bytesSinceLast": chunk_size, |
| | } |
| | ) |
| | |
| |
|
| | r = get_session().post( |
| | completion_url, |
| | json={ |
| | "oid": oid, |
| | "parts": parts, |
| | }, |
| | ) |
| | hf_raise_for_status(r) |
| |
|
| | write_msg({"event": "complete", "oid": oid}) |
| |
|