import os import json import sys import time import traceback from typing import Dict, Any from tqdm import tqdm import torch import torch.multiprocessing as mp from transformers import BertTokenizer worker_model_objects: Dict[str, Any] = {} def init_worker(model_path: str, bert_path: str, humanomni_project_path: str, device: str): global worker_model_objects if humanomni_project_path and humanomni_project_path not in sys.path: sys.path.append(humanomni_project_path) try: from humanomni import model_init, mm_infer from humanomni.utils import disable_torch_init except ImportError: print(f"[Worker PID: {os.getpid()}] ERROR: Failed to import HumanOmni. Ensure the humanomni_path is set correctly.", file=sys.stderr) return disable_torch_init() model, processor, tokenizer = model_init(model_path, device=device) bert_tokenizer = BertTokenizer.from_pretrained(bert_path) worker_model_objects = { "model": model, "processor": processor, "tokenizer": tokenizer, "bert_tokenizer": bert_tokenizer, "mm_infer": mm_infer, } def get_media_type(file_path: str) -> str: ext = os.path.splitext(file_path)[1].lower() if ext in ['.mp4', '.avi', '.mov', '.mkv', '.webm']: return 'video' elif ext in ['.jpg', '.jpeg', '.png', '.bmp', '.gif', '.webp']: return 'image' else: return 'unknown' def process_single_sample(media_full_path: str, prompt_text: str) -> str: global worker_model_objects try: model = worker_model_objects['model'] processor = worker_model_objects['processor'] tokenizer = worker_model_objects['tokenizer'] bert_tokenizer = worker_model_objects['bert_tokenizer'] mm_infer = worker_model_objects['mm_infer'] media_type = get_media_type(media_full_path) if media_type == 'unknown': raise ValueError(f"Unsupported media type for file: {media_full_path}") clean_prompt = prompt_text.replace("", "").replace("