import os import json import time import random from typing import Dict, Any, List, Optional, Tuple from openai import OpenAI import multiprocessing as mp # ------------------------- # IO helpers # ------------------------- def load_json(file_path: str): with open(file_path, "r", encoding="utf-8") as f: return json.load(f) def append_jsonl(path: str, obj: Dict[str, Any]): os.makedirs(os.path.dirname(path), exist_ok=True) with open(path, "a", encoding="utf-8") as f: f.write(json.dumps(obj, ensure_ascii=False) + "\n") def load_done_keys(jsonl_path: str): done = set() if not os.path.exists(jsonl_path): return done with open(jsonl_path, "r", encoding="utf-8") as f: for line in f: line = line.strip() if not line: continue try: obj = json.loads(line) k = obj.get("key") if k is not None: done.add(k) except Exception: continue return done # ------------------------- # LLM helpers # ------------------------- def safe_json_parse(text: str) -> Optional[Dict[str, Any]]: text = (text or "").strip() if text.startswith("```"): text = text.strip("`").strip() if text.lower().startswith("json"): text = text[4:].strip() try: return json.loads(text) except Exception: return None def call_chat_json( client: OpenAI, model: str, system: str, user: str, max_tokens: int, temperature: float, top_p: float = 0.9, max_retries: int = 4, ) -> Dict[str, Any]: last_err = None for attempt in range(1, max_retries + 1): try: resp = client.chat.completions.create( model=model, messages=[ {"role": "system", "content": system}, {"role": "user", "content": user}, ], temperature=temperature, top_p=top_p, max_tokens=max_tokens, ) text = resp.choices[0].message.content obj = safe_json_parse(text) if obj is None: raise ValueError(f"JSON parse failed. Raw: {text[:200]}...") return obj except Exception as e: last_err = str(e) time.sleep(1.0 * attempt) raise RuntimeError(f"call_chat_json failed after {max_retries} retries. Last error: {last_err}") def normalize_text(s: str) -> str: return " ".join((s or "").lower().split()) # ------------------------- # Binary question detection # ------------------------- _BINARY_ANS = {"yes", "no", "true", "false"} _AUX_START = { "is", "are", "was", "were", "do", "does", "did", "can", "could", "will", "would", "should", "has", "have", "had", "may", "might", "must", } def is_binary_qa(question: str, gt_answer: str) -> bool: a = normalize_text(gt_answer) if a in _BINARY_ANS: return True q = (question or "").strip().lower() if q.endswith("?"): first = q.split()[0] if q.split() else "" if first in _AUX_START: return True return False # ------------------------- # Round 1: Q/A -> correct statement # ------------------------- SYSTEM_STATEMENT = """Convert a VQA (question + ground-truth answer) into ONE declarative sentence. Return STRICT JSON: {"statement": "..."} only. Rules: - Must be accurate given the answer. - Natural, concise, factual. - Preserve referenced entity labels if any (e.g., "chair A"). - JSON only. """ USER_STATEMENT = """Question: {question} Ground-truth answer: {answer} Return JSON only. """ def gen_statement(client: OpenAI, model: str, question: str, answer: str) -> str: obj = call_chat_json( client=client, model=model, system=SYSTEM_STATEMENT, user=USER_STATEMENT.format(question=question, answer=answer), max_tokens=128, temperature=0.3, ) st = obj.get("statement", "") if not isinstance(st, str) or not st.strip(): raise ValueError(f"Bad statement: {obj}") return st.strip() # ------------------------- # Round 2: correct statement -> wrong options (N=9 default, N=1 for binary) # ------------------------- SYSTEM_WRONG_9 = """ You generate distractors for a ground-truth statement. OUTPUT MUST BE STRICT JSON ONLY: {{"distractors": ["", "", "", "", "", "", "", "", ""]}} Rules: - Exactly 9 strings, all unique. - Each is a single declarative sentence. - Each must be WRONG given the ground-truth answer. - Preserve the same main subject/entity (same instance label if present). - The distractors should be diverse, non-ambiguous, realistic. The difference between ground-truth and distractors should be VERY clear. - No markdown, no extra keys. """ USER_WRONG_9 = """INPUT Question: {question} Ground-truth answer: {answer} Ground-truth statement: {statement} TASK Generate 9 WRONG but plausible alternative declarative statements. REMINDER Output JSON only with the exact schema: {{"distractors": ["...","...","...","...","...","...","...","...","..."]}} """ SYSTEM_WRONG_1 = """ You generate distractors for a ground-truth statement. OUTPUT MUST BE STRICT JSON ONLY: {{"distractors": [""]}} Rules: - Exactly 1 string. - Must be a single declarative sentence. - Must be WRONG given the ground-truth answer. - Preserve the same main subject/entity (same instance label if present). - Make the wrong answer VERY clear (not ambiguous). - No markdown, no extra keys. """ USER_WRONG_1 = """INPUT Question: {question} Ground-truth answer: {answer} Ground-truth statement: {statement} TASK Generate 1 WRONG but plausible alternative declarative statement. REMINDER Output JSON only with the exact schema: {{"distractors": ["..."]}} """ def gen_wrong_sentences( client: OpenAI, model: str, question: str, answer: str, statement: str, n: int = 9, max_rounds: int = 5, ) -> List[str]: target = n collected: List[str] = [] seen = set([normalize_text(statement)]) # disallow matching correct if n == 1: system = SYSTEM_WRONG_1 user_tmpl = USER_WRONG_1 max_tokens = 256 temperature = 0.7 else: system = SYSTEM_WRONG_9 user_tmpl = USER_WRONG_9 max_tokens = 512 temperature = 0.85 for _round in range(max_rounds): obj = call_chat_json( client=client, model=model, system=system, user=user_tmpl.format(question=question, answer=answer, statement=statement), max_tokens=max_tokens, temperature=temperature, ) ds = obj.get("distractors", None) if not isinstance(ds, list): continue for x in ds: if not isinstance(x, str): continue x = x.strip() if not x: continue nx = normalize_text(x) if nx in seen: continue seen.add(nx) collected.append(x) if len(collected) >= target: return collected[:target] raise ValueError(f"Could not collect {target} unique distractors; got {len(collected)}") # ------------------------- # Build final options # ------------------------- def build_mcq(correct: str, wrongs: List[str], seed: Optional[int] = None): options = [correct] + list(wrongs) if seed is not None: rnd = random.Random(seed) rnd.shuffle(options) else: random.shuffle(options) label = options.index(correct) return options, label # ------------------------- # Multiprocessing worker # ------------------------- _WORKER_CLIENT = None _WORKER_MODEL = None def _init_worker(base_url: str, model_name: str, timeout: int = 3600): """Initializer runs once per process.""" global _WORKER_CLIENT, _WORKER_MODEL _WORKER_MODEL = model_name _WORKER_CLIENT = OpenAI(api_key="EMPTY", base_url=base_url, timeout=timeout) def _process_one(args: Tuple[str, int, str, str, str]) -> Dict[str, Any]: """ Args: key, idx, scan_id, question, gt_answer Returns: {"ok": True, "out": {...}} or {"ok": False, "err": {...}} """ key, idx, scan_id, question, gt_answer = args try: correct_stmt = gen_statement(_WORKER_CLIENT, _WORKER_MODEL, question, gt_answer) n_wrong = 1 if is_binary_qa(question, gt_answer) else 9 wrongs = gen_wrong_sentences( _WORKER_CLIENT, _WORKER_MODEL, question, gt_answer, correct_stmt, n=n_wrong, max_rounds=6 if n_wrong == 9 else 4, ) seed = abs(hash(key)) % (2**32) options, label = build_mcq(correct_stmt, wrongs, seed=seed) out = { "key": key, "scan_id": scan_id, "question": question, "gt_answer": gt_answer, "correct": correct_stmt, "options": options, # length 2 for binary, length 10 otherwise "label": label, "is_binary": (n_wrong == 1), } return {"ok": True, "out": out} except Exception as e: return {"ok": False, "err": { "key": key, "scan_id": scan_id, "question": question, "gt_answer": gt_answer, "error": str(e), }} # ------------------------- # Main # ------------------------- def main(): base_url = "http://lrc-alpha-sg-gpu06:22001/v1" model_name = "Qwen/Qwen3-VL-8B-Instruct" in_path = "/home/m50048399/transfered/ye_project/PointMapVerse/existing_datasets/ScanNet/annotations/qa/ScanQA_v1.0_val.json" out_jsonl = "/home/m50048399/transfered/ye_project/PointMapVerse/derived/scanqa_sentence_mcq_val.jsonl" err_jsonl = out_jsonl + ".errors.jsonl" data = load_json(in_path) done = load_done_keys(out_jsonl) print(f"Loaded {len(data)} items. Already done: {len(done)}") # Build task list tasks = [] for idx, item in enumerate(data): scan_id = item.get("scene_id", "") # keep your current field question = (item.get("question") or "").strip() answers = item.get("answers") or [] gt_answer = (answers[0] if answers else "").strip() if not question or not gt_answer: continue key = f"{scan_id}::{idx}" if key in done: continue tasks.append((key, idx, scan_id, question, gt_answer)) print(f"To process: {len(tasks)}") # Multiprocessing: 8 workers ctx = mp.get_context("spawn") with ctx.Pool( processes=8, initializer=_init_worker, initargs=(base_url, model_name, 3600), maxtasksperchild=50, ) as pool: processed = 0 ok_cnt = 0 err_cnt = 0 for res in pool.imap_unordered(_process_one, tasks, chunksize=4): processed += 1 if res["ok"]: append_jsonl(out_jsonl, res["out"]) ok_cnt += 1 else: append_jsonl(err_jsonl, res["err"]) err_cnt += 1 if processed % 100 == 0: print(f"Finished {processed}/{len(tasks)} | ok={ok_cnt} err={err_cnt}") print("Done.") if __name__ == "__main__": main()