flashtrace / exp /exp2 /map_math_mine_to_exp2_cache.py
wenbopan's picture
Sync FlashTrace package from GitHub
55b60a8
#!/usr/bin/env python3
"""Prepare data/math_mine.json into an exp2 cached JSONL dataset.
This script supports two modes:
- map (offline): convert GSM8K-style math examples:
{"question": "...", "answer": "... #### 18"}
into exp2's cached JSONL format (one JSON object per line).
- resample (online): resample targets like exp/exp2/sample_and_filter.py:
call a chat completion API to generate "<thinking> + final \\box{} answer",
judge the boxed answer against the reference answer extracted from the raw
GSM8K-style entry, and write only judge=True samples.
In both modes, exp2 expects token-level spans (NOT character spans):
- indices_to_explain: [start_tok, end_tok] (generation-token indices, closed interval)
- sink_span/thinking_span: token spans over tokenizer(target, add_special_tokens=False)
"""
from __future__ import annotations
import argparse
import json
import os
import sys
import time
import urllib.error
import urllib.request
from dataclasses import asdict
from pathlib import Path
from typing import Any, Dict, List, Optional, Tuple
from transformers import AutoTokenizer
from tqdm import tqdm
REPO_ROOT = Path(__file__).resolve().parents[2]
if str(REPO_ROOT) not in sys.path:
sys.path.insert(0, str(REPO_ROOT))
from exp.exp2.dataset_utils import CachedExample, attach_spans_from_answer, split_boxed_generation # noqa: E402
class RateLimitError(RuntimeError):
"""Raised when API returns 429; carries a suggested wait time."""
def __init__(self, wait_seconds: float, detail: str) -> None:
super().__init__(detail)
self.wait_seconds = wait_seconds
GEN_SYSTEM_PROMPT = (
"You are a reasoning assistant. "
"Before answering, engage in an chain of thought. "
"Process this freely and naturally without using specific headers or strict formatting. "
"When you reach the conclusion, wrap the entire final sentence containing the answer inside \\box{}. "
"Ensure the box wraps the **sentence** that naturally delivers the answer. DO NOT rewrite the answer word for the box separately."
)
JUDGE_SYSTEM_PROMPT = (
"You verify whether the model's boxed answer matches the reference answer. "
"Reply strictly with True or False and nothing else."
)
def call_chat_api(
api_base: str,
api_key: str,
model: str,
messages: List[Dict[str, str]],
*,
timeout: int,
max_tokens: int,
temperature: float,
cache_ttl: int,
cache_namespace: Optional[str],
rate_limit_delay: Optional[float] = None,
) -> str:
"""Minimal OpenAI-compatible chat.completions client (no external deps)."""
url = api_base.rstrip("/") + "/chat/completions"
payload: Dict[str, Any] = {
"model": model,
"messages": messages,
"max_tokens": max_tokens,
"temperature": temperature,
}
if cache_ttl > 0:
cache_obj: Dict[str, Any] = {"ttl": cache_ttl}
if cache_namespace:
cache_obj["namespace"] = cache_namespace
payload["cache"] = cache_obj
data = json.dumps(payload).encode("utf-8")
headers = {"Content-Type": "application/json"}
if api_key:
headers["Authorization"] = f"Bearer {api_key}"
req = urllib.request.Request(url, data=data, headers=headers, method="POST")
opener = urllib.request.build_opener(urllib.request.ProxyHandler({}))
try:
with opener.open(req, timeout=timeout) as resp:
resp_bytes = resp.read()
except urllib.error.HTTPError as e:
detail = e.read().decode("utf-8", errors="ignore") if hasattr(e, "read") else ""
if e.code == 429:
retry_after = None
if hasattr(e, "headers") and e.headers:
retry_after_header = e.headers.get("Retry-After")
if retry_after_header:
try:
retry_after = float(retry_after_header)
except ValueError:
retry_after = None
wait = retry_after or rate_limit_delay or 5.0
raise RateLimitError(wait, f"API HTTP 429: {detail}") from e
raise RuntimeError(f"API HTTP error {e.code}: {detail}") from e
except urllib.error.URLError as e:
raise RuntimeError(f"API request failed: {e}") from e
try:
response = json.loads(resp_bytes.decode("utf-8"))
except json.JSONDecodeError as e:
raise RuntimeError(f"Failed to decode API response: {resp_bytes!r}") from e
choices = response.get("choices", [])
if not choices:
raise RuntimeError(f"Empty choices from API: {response}")
content = choices[0].get("message", {}).get("content", "")
if not content:
raise RuntimeError(f"Empty content from API: {response}")
return content.strip()
def build_gen_messages(prompt: str) -> List[Dict[str, str]]:
return [
{"role": "system", "content": GEN_SYSTEM_PROMPT},
{"role": "user", "content": prompt},
]
def build_judge_messages(reference_answer: str, candidate_answer: str) -> List[Dict[str, str]]:
user = (
"Decide if the model's boxed answer matches the reference answer.\n"
f"Reference answer: {reference_answer}\n"
f"Model boxed answer (only the content inside \\box{{}}): {candidate_answer}\n"
"Output only True if they are semantically consistent; otherwise output False."
)
return [
{"role": "system", "content": JUDGE_SYSTEM_PROMPT},
{"role": "user", "content": user},
]
def parse_bool(text: str) -> bool:
first = text.strip().splitlines()[0].strip().lower()
if first in {"true", "yes"}:
return True
if first in {"false", "no"}:
return False
# fallback: check substring
if "true" in first and "false" not in first:
return True
if "false" in first:
return False
raise ValueError(f"Cannot parse boolean from: {text!r}")
def _load_tokenizer(tokenizer_model: str):
tok_path = Path(tokenizer_model)
if tok_path.exists():
tokenizer = AutoTokenizer.from_pretrained(tok_path.as_posix(), local_files_only=True)
else:
tokenizer = AutoTokenizer.from_pretrained(tokenizer_model)
if tokenizer.pad_token is None and tokenizer.eos_token is not None:
tokenizer.pad_token = tokenizer.eos_token
return tokenizer
def _split_gsm8k_answer(answer: str) -> Optional[Tuple[str, str]]:
"""Return (thinking_text, final_answer) parsed from GSM8K `answer`."""
text = (answer or "").strip()
if not text:
return None
if "####" not in text:
return None
thinking, final = text.rsplit("####", 1)
thinking = thinking.strip()
final = final.strip()
if not final:
return None
return thinking, final
def _is_token_span(span: Any) -> bool:
return isinstance(span, list) and len(span) == 2 and all(isinstance(x, int) for x in span)
def _build_cached_example(
*,
question: str,
answer: str,
tokenizer,
example_idx: int,
source_path: str,
) -> Optional[CachedExample]:
parsed = _split_gsm8k_answer(answer)
if parsed is None:
return None
thinking_text, final_answer = parsed
prompt = question.strip()
target = f"{thinking_text}\n{final_answer}" if thinking_text else final_answer
example = CachedExample(
prompt=prompt,
target=target,
indices_to_explain=None,
attr_mask_indices=None,
sink_span=None,
thinking_span=None,
metadata={
"dataset": "math_mine",
"source_path": source_path,
"example_idx": int(example_idx),
"raw_question": question,
"raw_answer": answer,
"reference_answer": final_answer,
"boxed_answer": final_answer,
},
)
example = attach_spans_from_answer(example, tokenizer, final_answer)
if not _is_token_span(example.sink_span):
return None
# exp2 requires token-level indices_to_explain=[start_tok,end_tok] (closed interval).
indices_to_explain = list(example.sink_span)
thinking_span = example.thinking_span
if thinking_span is not None and _is_token_span(thinking_span) and indices_to_explain[0] == 0:
# No room for "thinking" tokens; avoid overlapping spans.
thinking_span = None
return CachedExample(
prompt=example.prompt,
target=example.target,
indices_to_explain=indices_to_explain,
attr_mask_indices=example.attr_mask_indices,
sink_span=indices_to_explain,
thinking_span=thinking_span,
metadata=example.metadata,
)
def _build_resampled_example(
*,
question: str,
raw_answer: str,
reference_answer: str,
generation: str,
tokenizer,
example_idx: int,
source_path: str,
judge_response: str,
generator_model: str,
judge_model: str,
) -> Optional[CachedExample]:
parsed = split_boxed_generation(generation)
if not parsed:
return None
thinking_text, _boxed_segment, boxed_answer = parsed
target_text = f"{thinking_text}\n{boxed_answer}" if thinking_text else boxed_answer
example = CachedExample(
prompt=question.strip(),
target=target_text,
indices_to_explain=None,
attr_mask_indices=None,
sink_span=None,
thinking_span=None,
metadata={
"dataset": "math_mine",
"source_path": source_path,
"example_idx": int(example_idx),
"raw_question": question,
"raw_answer": raw_answer,
"reference_answer": reference_answer,
"judge_response": judge_response,
"generator_model": generator_model,
"judge_model": judge_model,
},
)
example = attach_spans_from_answer(example, tokenizer, boxed_answer)
if not _is_token_span(example.sink_span):
return None
indices_to_explain = list(example.sink_span)
return CachedExample(
prompt=example.prompt,
target=example.target,
indices_to_explain=indices_to_explain,
attr_mask_indices=example.attr_mask_indices,
sink_span=indices_to_explain,
thinking_span=example.thinking_span,
metadata=example.metadata,
)
def _write_jsonl(path: Path, *, examples) -> int:
path.parent.mkdir(parents=True, exist_ok=True)
count = 0
with path.open("w", encoding="utf-8") as f:
for ex in examples:
f.write(json.dumps(asdict(ex), ensure_ascii=False) + "\n")
count += 1
return count
def main() -> None:
ap = argparse.ArgumentParser("Prepare data/math_mine.json for exp2 cached JSONL.")
ap.add_argument("--in_json", type=str, default="data/math_mine.json")
ap.add_argument("--out_jsonl", type=str, default="exp/exp2/data/math.jsonl")
ap.add_argument(
"--tokenizer_model",
type=str,
required=True,
help="Tokenizer name or local path; must match the tokenizer used in exp2 attribution.",
)
ap.add_argument(
"--mode",
type=str,
choices=["map", "resample"],
default="map",
help="map=offline mapping from GSM8K answers; resample=generate+judge like exp/exp2/sample_and_filter.py.",
)
# Resample (online) options (kept compatible with exp/exp2/sample_and_filter.py).
ap.add_argument("--max_examples", type=int, default=100, help="Number of judge=True examples to keep (resample mode).")
ap.add_argument("--seed", type=int, default=42, help="Shuffle seed (only used with --shuffle).")
ap.add_argument("--shuffle", action="store_true", help="Shuffle examples before attempting (resample mode).")
ap.add_argument("--api_base", type=str, default="http://localhost:4000/v1", help="Chat API base URL.")
ap.add_argument("--api_key", type=str, default=None, help="API key; defaults to FLASHTRACE_API_KEY/OPENAI_API_KEY.")
ap.add_argument("--generator_model", type=str, default="qwen3-235b-a22b-2507")
ap.add_argument("--judge_model", type=str, default="deepseek-v3-1-terminus")
ap.add_argument("--api_timeout", type=int, default=300)
ap.add_argument("--api_max_tokens", type=int, default=8192)
ap.add_argument("--api_temperature", type=float, default=0.0)
ap.add_argument("--api_cache_ttl", type=int, default=600)
ap.add_argument("--api_cache_namespace", type=str, default="flashtrace-exp2")
ap.add_argument("--retry_delay", type=float, default=2.0)
ap.add_argument("--retries", type=int, default=2, help="Additional retries on API failure.")
ap.add_argument("--request_interval", type=float, default=1.0, help="Sleep seconds between generation calls.")
ap.add_argument("--judge_interval", type=float, default=1.0, help="Sleep seconds between judge calls.")
ap.add_argument("--rate_limit_delay", type=float, default=5.0, help="Seconds to wait on HTTP 429 before retrying.")
args = ap.parse_args()
in_path = Path(args.in_json)
out_path = Path(args.out_jsonl)
tokenizer = _load_tokenizer(args.tokenizer_model)
raw = json.loads(in_path.read_text(encoding="utf-8"))
if not isinstance(raw, list):
raise SystemExit(f"Expected a JSON array in {in_path}, got {type(raw).__name__}.")
source_total = len(raw)
total = 0
kept = 0
skipped_empty_q = 0
skipped_empty_a = 0
skipped_parse = 0
skipped_span = 0
examples = []
if args.mode == "map":
attempted = None
skipped_format = None
judged_false = None
for idx, item in enumerate(raw):
total += 1
if not isinstance(item, dict):
skipped_parse += 1
continue
question = str(item.get("question") or "")
answer = str(item.get("answer") or "")
if not question.strip():
skipped_empty_q += 1
continue
if not answer.strip():
skipped_empty_a += 1
continue
ex = _build_cached_example(
question=question,
answer=answer,
tokenizer=tokenizer,
example_idx=idx,
source_path=str(in_path),
)
if ex is None:
# distinguish parse-vs-span failure
parsed = _split_gsm8k_answer(answer)
if parsed is None:
skipped_parse += 1
else:
skipped_span += 1
continue
examples.append(ex)
kept += 1
else:
api_key = args.api_key or os.environ.get("FLASHTRACE_API_KEY") or os.environ.get("OPENAI_API_KEY")
if not api_key:
raise SystemExit("resample mode requires --api_key or FLASHTRACE_API_KEY/OPENAI_API_KEY.")
attempted = 0
skipped_format = 0
judged_false = 0
indices = list(range(len(raw)))
if bool(args.shuffle):
import random
rnd = random.Random(int(args.seed))
rnd.shuffle(indices)
kept_bar = tqdm(total=int(args.max_examples), desc="Kept (judge=True)", position=1, leave=False)
for loop_idx in tqdm(indices, total=len(indices), desc="Resampling"):
if kept >= int(args.max_examples):
break
total += 1
item = raw[loop_idx]
if not isinstance(item, dict):
skipped_parse += 1
continue
question = str(item.get("question") or "")
answer = str(item.get("answer") or "")
if not question.strip():
skipped_empty_q += 1
continue
if not answer.strip():
skipped_empty_a += 1
continue
parsed = _split_gsm8k_answer(answer)
if parsed is None:
skipped_parse += 1
continue
_ref_thinking, reference_answer = parsed
attempted += 1
gen_messages = build_gen_messages(question.strip())
# Step 1: generation
for attempt in range(int(args.retries) + 1):
try:
generation = call_chat_api(
str(args.api_base),
str(api_key),
str(args.generator_model),
gen_messages,
timeout=int(args.api_timeout),
max_tokens=int(args.api_max_tokens),
temperature=float(args.api_temperature),
cache_ttl=int(args.api_cache_ttl),
cache_namespace=str(args.api_cache_namespace) if args.api_cache_namespace else None,
rate_limit_delay=float(args.rate_limit_delay) if args.rate_limit_delay is not None else None,
)
break
except RateLimitError as e:
if attempt >= int(args.retries):
raise
time.sleep(float(e.wait_seconds))
except Exception: # noqa: BLE001
if attempt >= int(args.retries):
raise
time.sleep(float(args.retry_delay))
if float(args.request_interval) > 0:
time.sleep(float(args.request_interval))
parsed_gen = split_boxed_generation(generation)
if not parsed_gen:
skipped_format += 1
print(f"[attempt={attempted}] skipped=format")
continue
thinking_text, _boxed_segment, boxed_answer = parsed_gen
judge_messages = build_judge_messages(reference_answer, boxed_answer)
ok = False
judge_resp = ""
for attempt in range(int(args.retries) + 1):
try:
judge_resp = call_chat_api(
str(args.api_base),
str(api_key),
str(args.judge_model),
judge_messages,
timeout=int(args.api_timeout),
max_tokens=64,
temperature=0.0,
cache_ttl=int(args.api_cache_ttl),
cache_namespace=str(args.api_cache_namespace) if args.api_cache_namespace else None,
rate_limit_delay=float(args.rate_limit_delay) if args.rate_limit_delay is not None else None,
)
ok = parse_bool(judge_resp)
break
except RateLimitError as e:
if attempt >= int(args.retries):
raise
time.sleep(float(e.wait_seconds))
except Exception: # noqa: BLE001
if attempt >= int(args.retries):
raise
time.sleep(float(args.retry_delay))
if float(args.judge_interval) > 0:
time.sleep(float(args.judge_interval))
if not ok:
judged_false += 1
print(f"[attempt={attempted}] judge=filtered")
continue
ex = _build_resampled_example(
question=question,
raw_answer=answer,
reference_answer=reference_answer,
generation=generation,
tokenizer=tokenizer,
example_idx=int(loop_idx),
source_path=str(in_path),
judge_response=judge_resp,
generator_model=str(args.generator_model),
judge_model=str(args.judge_model),
)
if ex is None:
skipped_span += 1
print(f"[attempt={attempted}] skipped=span")
continue
examples.append(ex)
kept += 1
kept_bar.update(1)
print(f"[attempt={attempted}] judge=kept")
kept_bar.close()
written = _write_jsonl(out_path, examples=examples)
if written != kept:
raise SystemExit(f"Internal error: written={written} != kept={kept}")
print(
json.dumps(
{
"in_json": str(in_path),
"out_jsonl": str(out_path),
"tokenizer_model": args.tokenizer_model,
"mode": str(args.mode),
"source_total": int(source_total),
"visited": total,
"kept": kept,
"skipped_empty_question": skipped_empty_q,
"skipped_empty_answer": skipped_empty_a,
"skipped_parse": skipped_parse,
"skipped_span": skipped_span,
"attempted": attempted,
"skipped_format": skipped_format,
"judged_false": judged_false,
"max_examples": int(args.max_examples) if str(args.mode) == "resample" else None,
"api_base": str(args.api_base) if str(args.mode) == "resample" else None,
"generator_model": str(args.generator_model) if str(args.mode) == "resample" else None,
"judge_model": str(args.judge_model) if str(args.mode) == "resample" else None,
},
ensure_ascii=False,
indent=2,
)
)
if __name__ == "__main__":
main()