Upload mcp_remote_daily_job.py with huggingface_hub
Browse files- mcp_remote_daily_job.py +3 -6
mcp_remote_daily_job.py
CHANGED
|
@@ -17,7 +17,6 @@ Usage:
|
|
| 17 |
"""
|
| 18 |
|
| 19 |
import sys
|
| 20 |
-
import os
|
| 21 |
import re
|
| 22 |
from datetime import datetime
|
| 23 |
from collections import defaultdict
|
|
@@ -25,14 +24,12 @@ from datasets import load_dataset
|
|
| 25 |
|
| 26 |
|
| 27 |
def main():
|
| 28 |
-
#
|
| 29 |
-
hf_token = os.environ.get("HF_TOKEN")
|
| 30 |
-
|
| 31 |
print(f"[{datetime.now().isoformat()}] Loading dataset from evalstate/hf-mcp-logs...", file=sys.stderr)
|
| 32 |
|
| 33 |
# Load the source dataset (streaming for large datasets)
|
| 34 |
-
#
|
| 35 |
-
ds = load_dataset('evalstate/hf-mcp-logs', 'sessions', streaming=True
|
| 36 |
sessions_ds = ds['sessions']
|
| 37 |
|
| 38 |
# Pattern to detect mcp-remote suffix (any version)
|
|
|
|
| 17 |
"""
|
| 18 |
|
| 19 |
import sys
|
|
|
|
| 20 |
import re
|
| 21 |
from datetime import datetime
|
| 22 |
from collections import defaultdict
|
|
|
|
| 24 |
|
| 25 |
|
| 26 |
def main():
|
| 27 |
+
# Don't pass token explicitly - let datasets library pick it up from HF_TOKEN env var
|
|
|
|
|
|
|
| 28 |
print(f"[{datetime.now().isoformat()}] Loading dataset from evalstate/hf-mcp-logs...", file=sys.stderr)
|
| 29 |
|
| 30 |
# Load the source dataset (streaming for large datasets)
|
| 31 |
+
# The HF_TOKEN is passed as a secret and will be picked up automatically
|
| 32 |
+
ds = load_dataset('evalstate/hf-mcp-logs', 'sessions', streaming=True)
|
| 33 |
sessions_ds = ds['sessions']
|
| 34 |
|
| 35 |
# Pattern to detect mcp-remote suffix (any version)
|