| #!/usr/bin/env bash |
| set -euo pipefail |
|
|
| SCRIPT_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)" |
| source "$SCRIPT_DIR/common.sh" |
| prepare_dataset_root |
|
|
| BUCKET="${BUCKET:-hf://buckets/lekdan/osu-everything}" |
| BUCKET_ID="${BUCKET_ID:-lekdan/osu-everything}" |
| STATE_DB="${STATE_DB:-.fetcher/state.db}" |
| REMOVE_COUNT="${REMOVE_COUNT:-1000}" |
| REMOVE_WORKERS="${REMOVE_WORKERS:-4}" |
| UPLOAD="${UPLOAD:-1}" |
| LATEST_REBUILD_WORKERS="${LATEST_REBUILD_WORKERS:-8}" |
|
|
| mkdir -p logs .scratch |
| LOG="logs/remove-maps-v1-$(date -u +%Y%m%dT%H%M%SZ).log" |
| SUMMARY=".scratch/remove-maps-v1-summary.json" |
| exec > >(tee -a "$LOG") 2>&1 |
|
|
| echo "started_at=$(date -u +%Y-%m-%dT%H:%M:%SZ)" |
| echo "source_root=${SOURCE_ROOT}" |
| echo "dataset_root=${DATASET_ROOT}" |
| echo "bucket=${BUCKET}" |
| echo "remove_count=${REMOVE_COUNT}" |
| echo "remove_workers=${REMOVE_WORKERS}" |
| echo "state_db=${STATE_DB}" |
| echo "upload=${UPLOAD}" |
|
|
| active="$(pgrep -af '[o]su_fetcher|[o]su_indexer|[i]ngest_osz|[r]ebuild_latest_snapshot|[c]ompact_metadata_v1|[u]pdate_maps_v1|[h]f( sync| buckets sync)' || true)" |
| if [ -n "$active" ]; then |
| echo "refusing to remove maps while another pipeline process is active:" >&2 |
| echo "$active" >&2 |
| exit 65 |
| fi |
|
|
| export HF_HUB_ENABLE_HF_TRANSFER=1 |
| export HF_XET_HIGH_PERFORMANCE=1 |
| export HF_XET_CLIENT_AC_INITIAL_UPLOAD_CONCURRENCY="${HF_XET_CLIENT_AC_INITIAL_UPLOAD_CONCURRENCY:-4}" |
| export HF_XET_CLIENT_AC_MAX_UPLOAD_CONCURRENCY="${HF_XET_CLIENT_AC_MAX_UPLOAD_CONCURRENCY:-8}" |
| export HF_XET_CLIENT_RETRY_MAX_DURATION="${HF_XET_CLIENT_RETRY_MAX_DURATION:-600}" |
|
|
| "$PYTHON" "$PY_SOURCE_ROOT/python/validate_compact_v1.py" \ |
| --repo-root "$PY_DATASET_ROOT" \ |
| --require-archive-file-match \ |
| --max-data-files 10000 |
|
|
| "$PYTHON" "$PY_SOURCE_ROOT/python/remove_maps_v1.py" \ |
| --repo-root "$PY_DATASET_ROOT" \ |
| --count "$REMOVE_COUNT" \ |
| --state-db "$STATE_DB" \ |
| --clear-enumerate-state \ |
| --workers "$REMOVE_WORKERS" \ |
| --summary-path "$SUMMARY" |
|
|
| "$PYTHON" "$PY_SOURCE_ROOT/python/rebuild_latest_snapshot.py" \ |
| --repo-root "$PY_DATASET_ROOT" \ |
| --workers "$LATEST_REBUILD_WORKERS" |
|
|
| "$PYTHON" "$PY_SOURCE_ROOT/python/validate_compact_v1.py" \ |
| --repo-root "$PY_DATASET_ROOT" \ |
| --require-archive-file-match \ |
| --max-data-files 10000 \ |
| --json |
|
|
| "$PYTHON" "$PY_SOURCE_ROOT/python/seed_fetcher_state.py" \ |
| --repo-root "$PY_DATASET_ROOT" \ |
| --state-db "$STATE_DB" \ |
| --checkpoint-only |
|
|
| if [ "$UPLOAD" = "1" ]; then |
| echo "syncing destructive archive deletes to bucket" |
| "$HF" sync archives "${BUCKET}/archives" --delete |
| "$HF" sync data "${BUCKET}/data" --delete |
| "$HF" sync schemas "${BUCKET}/schemas" --delete |
|
|
| rm -rf .scratch/fetcher-state-upload |
| mkdir -p .scratch/fetcher-state-upload |
| cp "$STATE_DB" .scratch/fetcher-state-upload/state.db |
| "$HF" sync .scratch/fetcher-state-upload "${BUCKET}/state/fetcher" --delete |
|
|
| "$HF" buckets info "$BUCKET_ID" |
| fi |
|
|
| echo "finished_at=$(date -u +%Y-%m-%dT%H:%M:%SZ)" |
|
|