| import time, os, json |
| from tqdm import tqdm |
| from curl_cffi import requests |
| import concurrent.futures |
| from pathlib import Path |
| import tarfile |
| import shutil |
| from huggingface_hub import HfApi |
|
|
| os.environ["HF_HUB_ENABLE_HF_TRANSFER"] = "1" |
|
|
| def main(): |
| danbooru_path = Path("danbooru") |
| selected_range = set(['%03d' % i for i in [400]]) |
| notexists, total = 0, 0 |
| notexistsid = {} |
| with open("posts.json", "r") as f: |
| bar = tqdm(desc="Indexing files", ascii=True, leave=False) |
| cache = {} |
| data_size = {} |
| for d in selected_range: |
| data = [] |
| fp = ('0' + d) |
| for file_path in (danbooru_path / fp).iterdir(): |
| if file_path.is_file(): |
| data.append(file_path.stem) |
| data_size[file_path.stem] = file_path.stat().st_size |
| bar.update(1) |
| cache[fp] = set(data) |
|
|
| bar = tqdm(desc="Checking files", total=6_857_737, ascii=True, leave=False) |
| for line in f: |
| post = json.loads(line) |
| file_id = post['id'] |
| cutoff = str(file_id)[-3:] |
| if cutoff not in selected_range: |
| bar.update(1) |
| continue |
| cutoff = '0' + cutoff |
| assert (danbooru_path / cutoff).exists(), f"{(danbooru_path / cutoff)} not exixts" |
| exists = str(file_id) in cache[cutoff] and (data_size[str(file_id)] == post["file_size"] or int(file_id) < 5_020_995) |
| total += 1 |
| if not exists and "file_url" in post: |
| notexists += 1 |
| if cutoff not in notexistsid: |
| notexistsid[cutoff] = [] |
| notexistsid[cutoff].append(( |
| file_id, cutoff, post["file_url"], |
| )) |
| |
| bar.update(1) |
| bar.set_postfix_str(f"not exists: {notexists}") |
|
|
| bar.close() |
| print(f"not exists: {notexists}, total: {total}") |
|
|
|
|
| with concurrent.futures.ThreadPoolExecutor(max_workers=4) as tar_executor: |
| for d in selected_range: |
| cut = '0' + d |
| if cut not in notexistsid: |
| tar_executor.submit(archive_and_upload, Path("danbooru") / cut, cut) |
|
|
| for key, group in notexistsid.items(): |
| keybar = tqdm(desc=f"Downloading files in key={key}", total=len(group), position=1, ascii=True, leave=False) |
| ok = False |
| while not ok: |
| with concurrent.futures.ThreadPoolExecutor(max_workers=3) as executor: |
| for file_id, cutoff, file_url in group: |
| executor.submit(download, file_id, cutoff, file_url, keybar) |
| ok = True |
| for file_id, cutoff, file_url in group: |
| suffix = Path(file_url).suffix |
| if file_url != "" and not Path(f"danbooru/{cutoff}/{file_id}{suffix}").is_file(): |
| ok = False |
| tar_executor.submit(archive_and_upload, Path(f"danbooru/{cutoff}"), cutoff) |
| print(f"Finished download group {cutoff}") |
| keybar.close() |
|
|
| def rm_tree(pth: Path): |
| for child in pth.iterdir(): |
| if child.is_file(): |
| child.unlink() |
| else: |
| rm_tree(child) |
| pth.rmdir() |
|
|
| def archive_and_upload(dirname, name): |
| tar_name = Path("danbooru-tars") / f"data-{name}.tar" |
| |
| if not os.path.isdir(dirname): |
| print("The specified directory does not exist.") |
| return |
| |
| |
| print(f"Creating {tar_name}") |
| with tarfile.open(tar_name, "w") as tar: |
| |
| for root, dirs, files in os.walk(dirname): |
| |
| for file in tqdm(sorted(files), desc=f"Creating {tar_name}", ascii=True): |
| full_path = os.path.join(root, file) |
| |
| tar.add(full_path, arcname=file) |
| |
| |
| rm_tree(dirname) |
| print(f"The directory {dirname} has been removed.") |
|
|
| api = HfApi() |
| print(api.upload_file( |
| path_or_fileobj=tar_name, |
| path_in_repo=f"original/data-{name}.tar", |
| repo_id="nyanko7/danbooru2023", |
| repo_type="dataset", |
| )) |
| Path(tar_name).unlink() |
|
|
| def download(idx, cutoff, file_url, bar): |
| suffix = Path(file_url).suffix |
| max_attempts = 3 |
|
|
| for attempt in range(max_attempts): |
| try: |
| r = requests.get(file_url, impersonate="chrome110", timeout=120) |
| if r.status_code == 200: |
| with open(f"danbooru/{cutoff}/{idx}{suffix}", "wb") as f: |
| f.write(r.content) |
| break |
| else: |
| print(f"Attempt {attempt+1} failed to download {file_url}: error {r.status_code}") |
| except Exception as e: |
| print(f"Attempt {attempt+1} failed to download {file_url}: error {e}") |
|
|
| time.sleep(1) |
| if attempt+1 == max_attempts: |
| print(f"Failed to download {file_url} after {max_attempts} attempts.") |
| bar.update(1) |
|
|
| if __name__ == "__main__": |
| main() |
|
|