| |
| import os, time, glob, datetime, json |
| from huggingface_hub import HfApi |
|
|
| TOKEN = 'HF_TOKEN_REDACTED' |
| REPO = 'OpenTransformer/web-crawl-2026' |
| STALE_SEC = 120 |
| POLL_SEC = 30 |
| MIN_SIZE = 50 * 1024 * 1024 |
| DIRS = ['/workspace/scraped_data_go', '/workspace/scraped_data', '/workspace/staging'] |
| UPLOADED_FILE = '/workspace/uploaded_fast.json' |
| api = HfApi(token=TOKEN) |
|
|
| def log(msg): |
| ts = datetime.datetime.utcnow().isoformat() |
| print(f'{ts} {msg}', flush=True) |
|
|
| def load_uploaded(): |
| try: |
| with open(UPLOADED_FILE) as f: |
| return set(json.load(f)) |
| except: |
| return set() |
|
|
| def save_uploaded(s): |
| with open(UPLOADED_FILE, 'w') as f: |
| json.dump(list(s), f) |
|
|
| def main(): |
| log('Fast upload daemon started (30s poll)') |
| uploaded = load_uploaded() |
| while True: |
| now = time.time() |
| for d in DIRS: |
| if not os.path.isdir(d): |
| continue |
| for f in glob.glob(os.path.join(d, '*.gz')): |
| basename = os.path.basename(f) |
| if basename in uploaded: |
| continue |
| sz = os.path.getsize(f) |
| age = now - os.path.getmtime(f) |
| if sz < MIN_SIZE or age < STALE_SEC: |
| continue |
| sz_mb = sz / (1024*1024) |
| ts = datetime.datetime.utcnow().strftime('%Y%m%d_%H%M%S') |
| stem = basename.replace('.jsonl.gz', '') |
| remote = 'crawl/combined/' + stem + '_' + ts + '.jsonl.gz' |
| log('Uploading ' + basename + ' (' + str(int(sz_mb)) + 'MB) -> ' + remote) |
| try: |
| api.upload_file( |
| path_or_fileobj=f, |
| path_in_repo=remote, |
| repo_id=REPO, |
| repo_type='dataset', |
| commit_message='Crawl data: ' + basename + ' (' + str(int(sz_mb)) + 'MB)', |
| ) |
| log('Upload complete! Deleting local file.') |
| uploaded.add(basename) |
| save_uploaded(uploaded) |
| try: |
| os.remove(f) |
| except: |
| pass |
| except Exception as e: |
| log('Upload FAILED: ' + str(e)) |
| time.sleep(POLL_SEC) |
|
|
| if __name__ == '__main__': |
| main() |
|
|