| """ |
| Prepares PES2O for release on the HugginFace hub. |
| |
| Author: Luca Soldaini (@soldni) |
| """ |
|
|
| import argparse |
| import json |
| from contextlib import ExitStack |
| from functools import partial |
| from multiprocessing import Manager, Pool, cpu_count, set_start_method |
| from pathlib import Path |
| from queue import Queue |
| from threading import Thread |
| from time import sleep |
| import re |
| from typing import List, Optional, Tuple |
|
|
| from smashed.utils import ( |
| MultiPath, |
| recursively_list_files, |
| ) |
| import smart_open |
| from tqdm import tqdm |
|
|
|
|
| def process_single( |
| io_paths: Tuple[List[MultiPath], MultiPath], |
| version: str, |
| pbar_queue: Optional[Queue] = None, |
| ): |
| all_src, dst = io_paths |
| docs_cnt = 0 |
|
|
| if dst.is_local: |
| Path(dst.as_str).parent.mkdir(parents=True, exist_ok=True) |
|
|
| with smart_open.open(dst.as_str, "wt") as out_stream: |
| for src in all_src: |
| with smart_open.open(src.as_str, "rt") as in_stream: |
|
|
| for line in in_stream: |
| data = json.loads(line) |
| data.pop("metadata", None) |
| data["source"] = ( |
| ("s2ag" if "dataset=s2ag" in src.as_str else "s2orc") + |
| '/' + |
| ("train" if "split=train" in src.as_str else "valid") |
| ) |
| data["version"] = version |
|
|
| out_stream.write(json.dumps(data) + "\n") |
| docs_cnt += 1 |
|
|
| if pbar_queue is not None and docs_cnt % 10000 == 0: |
| pbar_queue.put((0, docs_cnt)) |
| docs_cnt = 0 |
|
|
| if pbar_queue is not None: |
| pbar_queue.put((1, 0)) |
|
|
| if pbar_queue is not None: |
| pbar_queue.put((0, docs_cnt)) |
|
|
|
|
| def threaded_progressbar(q: Queue, timeout: float, total_files: Optional[int] = None): |
| with ExitStack() as stack: |
| files_pbar = stack.enter_context( |
| tqdm(desc=" Files", unit="files", position=0, total=total_files) |
| ) |
| docs_pbar = stack.enter_context( |
| tqdm(desc=" Docs", unit=" docs", position=1, unit_scale=True) |
| ) |
|
|
| while True: |
| item = q.get() |
| if item is None: |
| break |
| else: |
| files, docs = item |
| files_pbar.update(files) |
| docs_pbar.update(docs) |
| sleep(timeout) |
|
|
|
|
| def main(): |
| ap = argparse.ArgumentParser() |
| ap.add_argument("src", type=str, help="Source path") |
| ap.add_argument("dst", type=str, help="Destination path") |
| ap.add_argument("--debug", default=False, help="Debug mode", action="store_true") |
| ap.add_argument( |
| "--parallel", type=int, default=cpu_count(), help="Number of parallel processes" |
| ) |
| ap.add_argument( |
| "-v", "--version", type=str, required=True, help="Version of the dataset" |
| ) |
| opts = ap.parse_args() |
|
|
| src = MultiPath.parse(opts.src) |
| dst = MultiPath.parse(opts.dst) |
|
|
| |
| grouped_src_paths = {} |
| for single_src in recursively_list_files(src): |
| single_src_dir, _ = single_src.rsplit("/", 1) |
| grouped_src_paths.setdefault(single_src_dir, []).append(single_src) |
|
|
| src_paths: List[List[MultiPath]] = [] |
| dst_paths: List[MultiPath] = [] |
| for dir_name, dir_values in grouped_src_paths.items(): |
| src_paths.append([MultiPath.parse(p) for p in dir_values]) |
| dir_path = MultiPath.parse(dir_name.replace("part_id=", "") + ".gz") |
| dst_path = dst / (diff) if len(diff := (dir_path - src)) > 0 else dst |
| dst_path.path = re.sub(r'dataset=(\w+)/', '\\1.', dst_path.path) |
| dst_path.path = re.sub(r'split=(\w+)/', '\\1.', dst_path.path) |
| dst_paths.append(dst_path) |
|
|
| if opts.debug: |
| with tqdm(total=len(src_paths)) as pbar: |
| for single_src, single_dst in zip(src_paths, dst_paths): |
| process_single((single_src, single_dst), version=opts.version) |
| pbar.update(1) |
|
|
| else: |
| set_start_method("spawn") |
|
|
| with Pool(processes=opts.parallel) as pool: |
| pbar_queue: Queue = (manager := Manager()).Queue() |
| pbar_thread = Thread( |
| target=threaded_progressbar, |
| args=(pbar_queue, 0.1, sum(len(p) for p in src_paths)), |
| daemon=True, |
| ) |
| pbar_thread.start() |
|
|
| for _ in pool.imap_unordered( |
| partial(process_single, pbar_queue=pbar_queue, version=opts.version), |
| tuple(zip(src_paths, dst_paths)), |
| ): |
| ... |
|
|
| pool.close() |
| pool.join() |
|
|
| pbar_queue.put(None) |
| pbar_thread.join() |
| manager.shutdown() |
|
|
|
|
| if __name__ == "__main__": |
| main() |
|
|