| |
| """ |
| Task 5: Spectral Type Classification (ASIB Benchmark) |
| |
| Three questions evaluated: |
| Q1: Does the spectrum contain H-alpha and H-beta emission lines? (all 700 samples) |
| Q2: Is this a Broad-Line AGN (BLAGN)? (500 samples: groups C1-C4, D) |
| Q3: What is the BPT classification? (400 samples: groups C1-C4) |
| """ |
|
|
| import argparse |
| import base64 |
| import csv |
| import json |
| import os |
| import pathlib |
| import re |
| import time |
| from typing import Optional, List, Dict |
|
|
| from dotenv import load_dotenv |
| load_dotenv(override=True) |
|
|
| from openai import OpenAI |
|
|
|
|
| |
| |
| |
|
|
| DATA_DIR = pathlib.Path(__file__).resolve().parent.parent.parent / "data" / "Task5_SpecType" |
|
|
| |
| Q1_GROUPS = {"A", "B", "C1", "C2", "C3", "C4", "D"} |
| Q2_GROUPS = {"C1", "C2", "C3", "C4", "D"} |
| Q3_GROUPS = {"C1", "C2", "C3", "C4"} |
|
|
| |
| Q1_TRUTH = {"A": "False", "B": "False", "C1": "True", "C2": "True", "C3": "True", "C4": "True", "D": "True"} |
| Q2_TRUTH = {"C1": "False", "C2": "False", "C3": "False", "C4": "False", "D": "True"} |
| Q3_TRUTH = {"C1": "Star-Forming", "C2": "Composite", "C3": "Seyfert", "C4": "LINER"} |
|
|
| GROUP_TO_DIR = { |
| "A": "Group_A_High_Z_Trap", |
| "B": "Group_B_Low_Z_Weak", |
| "C1": "Group_C1_BPT_SF", |
| "C2": "Group_C2_BPT_Composite", |
| "C3": "Group_C3_BPT_Seyfert", |
| "C4": "Group_C4_BPT_LINER", |
| "D": "Group_D_Broad_Line_AGN", |
| } |
|
|
|
|
| |
| |
| |
|
|
| def get_client(model: str) -> OpenAI: |
| """Create OpenAI-compatible client based on model name. |
| |
| Requires environment variables: |
| - OPENAI_API_KEY / OPENAI_BASE_URL for OpenAI/compatible models |
| - CLAUDE_API_KEY for Claude models |
| - GROK_API_KEY for Grok models |
| - QWEN_API_KEY for Qwen models |
| - INTERN_API_KEY for InternVL models |
| """ |
| api_key = os.getenv("OPENAI_API_KEY") |
| base_url = os.getenv("OPENAI_BASE_URL") |
|
|
| if "intern" in model.lower(): |
| api_key = os.getenv("INTERN_API_KEY") |
| base_url = os.getenv("INTERN_BASE_URL") |
| elif "qwen" in model.lower(): |
| api_key = os.getenv("QWEN_API_KEY") |
| base_url = os.getenv("QWEN_BASE_URL") |
| elif "grok" in model.lower(): |
| api_key = os.getenv("GROK_API_KEY") |
| elif "claude" in model.lower(): |
| api_key = os.getenv("CLAUDE_API_KEY") |
|
|
| return OpenAI(api_key=api_key, base_url=base_url) |
|
|
|
|
| |
| |
| |
|
|
| def encode_image(path: pathlib.Path) -> str: |
| with open(path, "rb") as f: |
| return base64.b64encode(f.read()).decode("utf-8") |
|
|
|
|
| |
| |
| |
|
|
| def load_samples(question: str, data_dir: pathlib.Path = DATA_DIR) -> List[Dict]: |
| csv_path = data_dir / "ASIB_v1_selection_with_snr.csv" |
| figures_root = data_dir / "figures" |
|
|
| eligible_groups = {"Q1": Q1_GROUPS, "Q2": Q2_GROUPS, "Q3": Q3_GROUPS}[question] |
| truth_map = {"Q1": Q1_TRUTH, "Q2": Q2_TRUTH, "Q3": Q3_TRUTH}[question] |
|
|
| samples = [] |
| with csv_path.open(newline="") as f: |
| reader = csv.DictReader(f) |
| for row in reader: |
| group = row["SUB_GROUP"].strip() |
| if group not in eligible_groups: |
| continue |
|
|
| target_id = row["TARGETID"].strip() |
| group_dir = GROUP_TO_DIR[group] |
| fig_path = figures_root / group_dir / f"spectrum_{target_id}.png" |
|
|
| if not fig_path.exists(): |
| continue |
|
|
| samples.append({ |
| "target_id": target_id, |
| "group": group, |
| "label": truth_map[group], |
| "image_path": fig_path, |
| "z_conf": float(row.get("Z_CONF", 0) or 0), |
| "class_label": row.get("CLASS_LABEL", "").strip(), |
| }) |
|
|
| return samples |
|
|
|
|
| |
| |
| |
|
|
| SYSTEM_PROMPT_Q1 = """**Task:** Analyze this optical spectrum of a galaxy to determine whether the spectrum contains BOTH H-alpha (rest wavelength 6563 Angstroms) and H-beta (rest wavelength 4861 Angstroms) emission lines. |
| |
| **Context:** |
| - The spectrum is shown in the observed frame |
| - H-alpha and H-beta may be redshifted out of the optical wavelength range or may be weak/non-detectable |
| - Emission lines appear as peaks above the continuum |
| |
| **Output requirements:** |
| - Respond with a JSON object in the following format: {"answer": "", "reason": ""} |
| - The "answer" field must be either: True or False |
| - Answer True if both lines are present, False otherwise |
| - The "reason" field should contain a brief explanation |
| - Do not include any text outside the JSON object |
| """ |
|
|
| SYSTEM_PROMPT_Q2 = """**Task:** Analyze this optical spectrum of a galaxy to determine whether this object is a Broad-Line AGN (BLAGN). |
| |
| **Context:** |
| - A BLAGN is characterized by broad emission lines (FWHM > 1000 km/s), particularly in H-alpha |
| - The broad line component is wider than typical narrow-line regions |
| - Type-1 AGN / Seyfert 1 classification |
| - The H-alpha feature could be blended with [N II] lines which may complicate the profile |
| - Look for asymmetric or broadened emission line profiles in the Balmer lines (H-alpha, H-beta) |
| |
| **Output requirements:** |
| - Respond with a JSON object in the following format: {"answer": "", "reason": ""} |
| - The "answer" field must be either: True or False |
| - Answer True if this is a BLAGN, False otherwise |
| - The "reason" field should contain a brief explanation |
| - Do not include any text outside the JSON object |
| """ |
|
|
| SYSTEM_PROMPT_Q3 = """**Task:** Analyze this optical spectrum of a galaxy and classify it using the BPT (Baldwin-Phillips-Terlevich) diagnostic diagram. |
| |
| **Context:** |
| The BPT diagram uses the line ratios: |
| - log([NII] 6584 / H-alpha) on the x-axis |
| - log([OIII] 5007 / H-beta) on the y-axis |
| |
| The four classification regions are: |
| 1. Star-Forming: Low [NII]/H-alpha and low [OIII]/H-beta (ionization dominated by young stars) |
| 2. Composite: Intermediate region between SF and AGN (mixed ionization sources) |
| 3. Seyfert: High [OIII]/H-beta, high [NII]/H-alpha (AGN-dominated ionization) |
| 4. LINER: High [NII]/H-alpha, low [OIII]/H-beta (low-ionization nuclear emission region) |
| |
| **Output requirements:** |
| - Respond with a JSON object in the following format: {"answer": "", "reason": ""} |
| - The "answer" field must be one of: Star-Forming, Composite, Seyfert, or LINER |
| - The "reason" field should contain a brief explanation |
| - Do not include any text outside the JSON object |
| """ |
|
|
| SYSTEM_PROMPT_Q1_WOGUIDE = """Does this spectrum contain both H-alpha and H-beta emission lines? |
| |
| Output requirements: |
| - Respond with a JSON object: {"answer": "", "reason": ""} |
| - The "answer" field must be either: True or False |
| - Do not include any text outside the JSON object |
| """ |
|
|
| SYSTEM_PROMPT_Q2_WOGUIDE = """Is this a Broad-Line AGN (BLAGN)? |
| |
| Output requirements: |
| - Respond with a JSON object: {"answer": "", "reason": ""} |
| - The "answer" field must be either: True or False |
| - Do not include any text outside the JSON object |
| """ |
|
|
| SYSTEM_PROMPT_Q3_WOGUIDE = """Classify this spectrum using the BPT diagram: Star-Forming, Composite, Seyfert, or LINER. |
| |
| Output requirements: |
| - Respond with a JSON object: {"answer": "", "reason": ""} |
| - The "answer" field must be one of: Star-Forming, Composite, Seyfert, or LINER |
| - Do not include any text outside the JSON object |
| """ |
|
|
| USER_TEXT_Q1 = "Does this spectrum contain both H-alpha and H-beta emission lines? Respond with JSON format." |
| USER_TEXT_Q2 = "Is this a Broad-Line AGN (BLAGN)? Respond with JSON format." |
| USER_TEXT_Q3 = "Classify this spectrum: Star-Forming, Composite, Seyfert, or LINER. Respond with JSON format." |
|
|
|
|
| def get_prompts(question: str, prompt_type: str = "guided"): |
| if prompt_type == "guided": |
| prompts = {"Q1": SYSTEM_PROMPT_Q1, "Q2": SYSTEM_PROMPT_Q2, "Q3": SYSTEM_PROMPT_Q3} |
| else: |
| prompts = {"Q1": SYSTEM_PROMPT_Q1_WOGUIDE, "Q2": SYSTEM_PROMPT_Q2_WOGUIDE, "Q3": SYSTEM_PROMPT_Q3_WOGUIDE} |
| user_texts = {"Q1": USER_TEXT_Q1, "Q2": USER_TEXT_Q2, "Q3": USER_TEXT_Q3} |
| return prompts[question], user_texts[question] |
|
|
|
|
| |
| |
| |
|
|
| def classify_image( |
| client: OpenAI, |
| image_path: pathlib.Path, |
| model: str, |
| system_prompt: str, |
| user_text: str, |
| max_completion_tokens: int, |
| ): |
| img_b64 = encode_image(image_path) |
|
|
| messages = [ |
| {"role": "system", "content": system_prompt}, |
| { |
| "role": "user", |
| "content": [ |
| {"type": "text", "text": user_text}, |
| { |
| "type": "image_url", |
| "image_url": { |
| "url": f"data:image/png;base64,{img_b64}", |
| "detail": "high", |
| }, |
| }, |
| ], |
| }, |
| ] |
|
|
| extra = {"enable_thinking": False} if "qwen" in model.lower() else {} |
| for attempt in range(5): |
| try: |
| response = client.chat.completions.create( |
| model=model, |
| messages=messages, |
| temperature=0, |
| max_completion_tokens=max_completion_tokens, |
| extra_body=extra if extra else None, |
| ) |
| return response |
| except Exception as e: |
| if attempt < 4: |
| wait = 2 ** attempt * 5 |
| print(f" Attempt {attempt+1} failed ({e}), retrying in {wait}s...") |
| time.sleep(wait) |
| else: |
| raise |
|
|
|
|
| |
| |
| |
|
|
| def parse_prediction(raw: str) -> dict: |
| cleaned = re.sub(r"```json\s*", "", raw) |
| cleaned = re.sub(r"```\s*", "", cleaned) |
| cleaned = cleaned.strip() |
| try: |
| return json.loads(cleaned) |
| except json.JSONDecodeError: |
| return {"answer": raw, "reason": ""} |
|
|
|
|
| def canonicalize_binary(value) -> str: |
| if isinstance(value, bool): |
| return "True" if value else "False" |
| val = str(value or "").strip().upper() |
| if val.startswith("TRUE") or val == "T": |
| return "True" |
| if val.startswith("FALSE") or val == "F": |
| return "False" |
| if "TRUE" in val and "FALSE" not in val: |
| return "True" |
| if "FALSE" in val and "TRUE" not in val: |
| return "False" |
| return "Unknown" |
|
|
|
|
| def canonicalize_bpt(value: str) -> str: |
| val = (value or "").strip().upper() |
| if "STAR-FORMING" in val or "STAR FORMING" in val or val == "SF": |
| return "Star-Forming" |
| if "COMPOSITE" in val: |
| return "Composite" |
| if "SEYFERT" in val: |
| return "Seyfert" |
| if "LINER" in val: |
| return "LINER" |
| return "Unknown" |
|
|
|
|
| def canonicalize_answer(question: str, value: str) -> str: |
| if question in ("Q1", "Q2"): |
| return canonicalize_binary(value) |
| return canonicalize_bpt(value) |
|
|
|
|
| |
| |
| |
|
|
| def run( |
| question: str, |
| model: str, |
| limit: Optional[int], |
| results_dir: pathlib.Path, |
| prompt_type: str, |
| max_completion_tokens: int, |
| resume: bool, |
| data_dir: pathlib.Path = DATA_DIR, |
| ) -> pathlib.Path: |
|
|
| client = get_client(model) |
| samples = load_samples(question, data_dir) |
| print(f"Loaded {len(samples)} samples for {question}") |
|
|
| results_dir.mkdir(parents=True, exist_ok=True) |
| out_path = results_dir / f"predictions-{question}-{prompt_type}-{model}.json" |
|
|
| system_prompt, user_text = get_prompts(question, prompt_type) |
|
|
| results = [] |
| processed_images = set() |
| if resume and out_path.exists(): |
| with out_path.open("r") as f: |
| results = json.load(f) |
| processed_images = {r["image"] for r in results} |
| print(f"Resuming from {len(results)} existing predictions") |
|
|
| correct = sum(r["correct"] for r in results) |
| total = len(results) |
|
|
| for i, sample in enumerate(samples): |
| if limit is not None and i >= limit: |
| break |
|
|
| image_path = sample["image_path"] |
| if str(image_path) in processed_images: |
| continue |
|
|
| label = sample["label"] |
|
|
| response = classify_image(client, image_path, model, system_prompt, user_text, max_completion_tokens) |
| content = response.choices[0].message.content |
| pred = parse_prediction(content) |
| answer = canonicalize_answer(question, pred.get("answer", "")) |
| is_correct = answer == label |
|
|
| total += 1 |
| correct += int(is_correct) |
|
|
| results.append({ |
| "image": str(image_path), |
| "target_id": sample["target_id"], |
| "group": sample["group"], |
| "label": label, |
| "prediction": pred, |
| "correct": int(is_correct), |
| "raw_response": response.model_dump(), |
| }) |
| print( |
| f"{question} {sample['target_id']} (group={sample['group']}): " |
| f"pred={answer} label={label} " |
| f"{'✓' if is_correct else '✗'}" |
| ) |
|
|
| with out_path.open("w") as f: |
| json.dump(results, f, indent=2) |
|
|
| if total > 0: |
| print(f"{question} Accuracy on {total} checked: {correct}/{total} = {correct/total:.2%}") |
|
|
| print(f"Saved predictions to {out_path}") |
| return out_path |
|
|
|
|
| |
| |
| |
|
|
| def parse_args() -> argparse.Namespace: |
| parser = argparse.ArgumentParser(description="Task5: Spectral Type Classification (ASIB)") |
| parser.add_argument("--question", choices=["Q1", "Q2", "Q3"], required=True) |
| parser.add_argument("--model", default="gpt-4o") |
| parser.add_argument("--prompt-type", choices=["guided", "woguide"], default="guided") |
| parser.add_argument("--limit", type=int, default=None) |
| parser.add_argument("--results-dir", type=pathlib.Path, default=pathlib.Path("./results")) |
| parser.add_argument("--max-completion-tokens", type=int, default=16384) |
| parser.add_argument("--resume", action="store_true") |
| return parser.parse_args() |
|
|
|
|
| if __name__ == "__main__": |
| args = parse_args() |
| run( |
| question=args.question, |
| model=args.model, |
| limit=args.limit, |
| results_dir=args.results_dir, |
| prompt_type=args.prompt_type, |
| max_completion_tokens=args.max_completion_tokens, |
| resume=args.resume, |
| ) |
|
|