| | |
| | """Build SFT conversation pairs from parsed transcript turns. |
| | |
| | Reads dataset/data/turns.jsonl and produces: |
| | - dataset/data/conversations.jsonl (ShareGPT multi-turn format) |
| | - dataset/data/sft_pairs.jsonl (individual instruction/response pairs) |
| | """ |
| |
|
| | import json |
| | import sys |
| | from collections import defaultdict |
| | from pathlib import Path |
| |
|
| | DATA_DIR = Path(__file__).resolve().parent.parent / "data" |
| | TURNS_FILE = DATA_DIR / "turns.jsonl" |
| | CONV_FILE = DATA_DIR / "conversations.jsonl" |
| | SFT_FILE = DATA_DIR / "sft_pairs.jsonl" |
| |
|
| | MIN_RESPONSE_LEN = 20 |
| | MAX_CONTEXT_TURNS = 3 |
| |
|
| |
|
| | def load_turns_by_file(path: Path) -> dict: |
| | """Load turns grouped by source_file, preserving turn_index order.""" |
| | groups = defaultdict(list) |
| | with open(path, encoding="utf-8") as f: |
| | for line in f: |
| | turn = json.loads(line) |
| | groups[turn["source_file"]].append(turn) |
| | |
| | for sf in groups: |
| | groups[sf].sort(key=lambda t: t["turn_index"]) |
| | return groups |
| |
|
| |
|
| | def merge_messages(turns: list[dict]) -> list[dict]: |
| | """Merge consecutive same-role turns into ShareGPT messages. |
| | |
| | Audrey → role "assistant", everyone else → role "human". |
| | Consecutive human turns from different speakers get speaker attribution. |
| | Consecutive assistant turns are concatenated. |
| | """ |
| | messages = [] |
| | for turn in turns: |
| | role = "assistant" if turn["is_audrey"] else "human" |
| | speaker = turn["speaker"] |
| | text = turn["text"].strip() |
| | if not text: |
| | continue |
| |
|
| | if messages and messages[-1]["role"] == role: |
| | |
| | prev = messages[-1] |
| | if role == "human": |
| | |
| | if prev.get("_multi_speaker") is False: |
| | |
| | prev["content"] = f"**{prev['speaker']}:** {prev['content']}" |
| | prev["_multi_speaker"] = True |
| | prev["content"] += f"\n\n**{speaker}:** {text}" |
| | prev["_multi_speaker"] = True |
| | |
| | if speaker not in prev["_speakers"]: |
| | prev["_speakers"].append(speaker) |
| | else: |
| | |
| | prev["content"] += f"\n\n{text}" |
| | else: |
| | messages.append({ |
| | "role": role, |
| | "speaker": speaker, |
| | "content": text, |
| | "_multi_speaker": False, |
| | "_speakers": [speaker], |
| | }) |
| |
|
| | |
| | for msg in messages: |
| | if msg["_multi_speaker"]: |
| | |
| | msg["speaker"] = ", ".join(msg["_speakers"]) |
| | del msg["_multi_speaker"] |
| | del msg["_speakers"] |
| |
|
| | return messages |
| |
|
| |
|
| | def build_conversation(source_file: str, turns: list[dict]) -> dict: |
| | """Build a ShareGPT-format conversation record.""" |
| | first = turns[0] |
| | conv_id = source_file.replace(".md", "") |
| | messages = merge_messages(turns) |
| |
|
| | |
| | participants = [] |
| | seen = set() |
| | for t in turns: |
| | if t["speaker"] not in seen: |
| | seen.add(t["speaker"]) |
| | participants.append(t["speaker"]) |
| |
|
| | audrey_word_count = sum( |
| | len(m["content"].split()) if first["language"] == "en" |
| | else len(m["content"]) |
| | for m in messages if m["role"] == "assistant" |
| | ) |
| |
|
| | return { |
| | "id": conv_id, |
| | "date": first["date"], |
| | "title": first["title"], |
| | "language": first["language"], |
| | "participants": participants, |
| | "messages": [{"role": m["role"], "speaker": m["speaker"], "content": m["content"]} for m in messages], |
| | "source_file": source_file, |
| | "num_turns": len(messages), |
| | "audrey_word_count": audrey_word_count, |
| | } |
| |
|
| |
|
| | def build_sft_pairs_from_conversation(conv: dict) -> list[dict]: |
| | """Extract SFT pairs from a multi-turn conversation. |
| | |
| | For each assistant turn, include up to MAX_CONTEXT_TURNS preceding turns |
| | as instruction context. |
| | """ |
| | pairs = [] |
| | messages = conv["messages"] |
| | pair_idx = 0 |
| |
|
| | for i, msg in enumerate(messages): |
| | if msg["role"] != "assistant": |
| | continue |
| | response = msg["content"] |
| | if len(response) < MIN_RESPONSE_LEN: |
| | continue |
| |
|
| | |
| | context_start = max(0, i - MAX_CONTEXT_TURNS) |
| | context_msgs = messages[context_start:i] |
| |
|
| | if not context_msgs: |
| | |
| | continue |
| |
|
| | |
| | instruction_parts = [] |
| | for cm in context_msgs: |
| | if cm["role"] == "human": |
| | instruction_parts.append(cm["content"]) |
| | else: |
| | |
| | instruction_parts.append(f"[Audrey Tang:] {cm['content']}") |
| | instruction = "\n\n".join(instruction_parts) |
| |
|
| | pairs.append({ |
| | "id": f"{conv['id']}/pair_{pair_idx:03d}", |
| | "date": conv["date"], |
| | "title": conv["title"], |
| | "language": conv["language"], |
| | "instruction": instruction, |
| | "response": response, |
| | "context_turns": len(context_msgs), |
| | "source_file": conv["source_file"], |
| | }) |
| | pair_idx += 1 |
| |
|
| | return pairs |
| |
|
| |
|
| | def build_sft_pairs_from_monologue(conv: dict) -> list[dict]: |
| | """Build SFT pairs from a monologue (Audrey-only single-speaker file). |
| | |
| | Split the text into paragraphs and use the title as instruction. |
| | """ |
| | pairs = [] |
| | messages = conv["messages"] |
| | if not messages: |
| | return pairs |
| |
|
| | |
| | full_text = "\n\n".join(m["content"] for m in messages) |
| |
|
| | |
| | paragraphs = [p.strip() for p in full_text.split("\n\n") if p.strip()] |
| | if not paragraphs: |
| | return pairs |
| |
|
| | |
| | if len(paragraphs) <= 3 or len(full_text) < 500: |
| | if len(full_text) >= MIN_RESPONSE_LEN: |
| | pairs.append({ |
| | "id": f"{conv['id']}/monologue_000", |
| | "date": conv["date"], |
| | "title": conv["title"], |
| | "language": conv["language"], |
| | "instruction": conv["title"], |
| | "response": full_text, |
| | "context_turns": 0, |
| | "source_file": conv["source_file"], |
| | }) |
| | return pairs |
| |
|
| | |
| | chunks = [] |
| | current_chunk = [] |
| | current_len = 0 |
| | for para in paragraphs: |
| | current_chunk.append(para) |
| | current_len += len(para) |
| | if current_len >= 400: |
| | chunks.append("\n\n".join(current_chunk)) |
| | current_chunk = [] |
| | current_len = 0 |
| | if current_chunk: |
| | chunks.append("\n\n".join(current_chunk)) |
| |
|
| | for idx, chunk in enumerate(chunks): |
| | if len(chunk) < MIN_RESPONSE_LEN: |
| | continue |
| | if idx == 0: |
| | instruction = conv["title"] |
| | else: |
| | instruction = f"{conv['title']} (continued)" |
| | pairs.append({ |
| | "id": f"{conv['id']}/monologue_{idx:03d}", |
| | "date": conv["date"], |
| | "title": conv["title"], |
| | "language": conv["language"], |
| | "instruction": instruction, |
| | "response": chunk, |
| | "context_turns": 0, |
| | "source_file": conv["source_file"], |
| | }) |
| |
|
| | return pairs |
| |
|
| |
|
| | def is_monologue(turns: list[dict]) -> bool: |
| | """Check if all turns are from a single speaker who is Audrey.""" |
| | speakers = {t["speaker"] for t in turns} |
| | return len(speakers) == 1 and turns[0]["is_audrey"] |
| |
|
| |
|
| | def main(): |
| | print(f"Reading turns from {TURNS_FILE}...") |
| | groups = load_turns_by_file(TURNS_FILE) |
| | print(f"Found {len(groups)} source files, {sum(len(v) for v in groups.values())} total turns") |
| |
|
| | conversations = [] |
| | all_sft_pairs = [] |
| | monologue_count = 0 |
| | no_audrey_count = 0 |
| | lang_counts = defaultdict(int) |
| | lang_sft_counts = defaultdict(int) |
| |
|
| | for source_file, turns in sorted(groups.items()): |
| | conv = build_conversation(source_file, turns) |
| | conversations.append(conv) |
| | lang_counts[conv["language"]] += 1 |
| |
|
| | has_audrey = any(t["is_audrey"] for t in turns) |
| |
|
| | if is_monologue(turns): |
| | monologue_count += 1 |
| | pairs = build_sft_pairs_from_monologue(conv) |
| | elif has_audrey: |
| | pairs = build_sft_pairs_from_conversation(conv) |
| | else: |
| | |
| | no_audrey_count += 1 |
| | pairs = [] |
| |
|
| | for p in pairs: |
| | lang_sft_counts[p["language"]] += 1 |
| | all_sft_pairs.extend(pairs) |
| |
|
| | |
| | print(f"\nWriting {len(conversations)} conversations to {CONV_FILE}...") |
| | with open(CONV_FILE, "w", encoding="utf-8") as f: |
| | for conv in conversations: |
| | f.write(json.dumps(conv, ensure_ascii=False) + "\n") |
| |
|
| | print(f"Writing {len(all_sft_pairs)} SFT pairs to {SFT_FILE}...") |
| | with open(SFT_FILE, "w", encoding="utf-8") as f: |
| | for pair in all_sft_pairs: |
| | f.write(json.dumps(pair, ensure_ascii=False) + "\n") |
| |
|
| | |
| | total_turns = sum(c["num_turns"] for c in conversations) |
| | avg_turns = total_turns / len(conversations) if conversations else 0 |
| | total_audrey_wc = sum(c["audrey_word_count"] for c in conversations) |
| |
|
| | print("\n" + "=" * 60) |
| | print("OUTPUT STATISTICS") |
| | print("=" * 60) |
| | print(f"Total conversations: {len(conversations)}") |
| | print(f" Monologues (Audrey only): {monologue_count}") |
| | print(f" No Audrey present: {no_audrey_count}") |
| | print(f" Multi-speaker w/ Audrey: {len(conversations) - monologue_count - no_audrey_count}") |
| | print(f"Total SFT pairs: {len(all_sft_pairs)}") |
| | print(f"Avg turns per conversation: {avg_turns:.1f}") |
| | print(f"Total Audrey word/char ct: {total_audrey_wc}") |
| | print() |
| | print("Language split (conversations):") |
| | for lang, count in sorted(lang_counts.items()): |
| | print(f" {lang}: {count}") |
| | print("Language split (SFT pairs):") |
| | for lang, count in sorted(lang_sft_counts.items()): |
| | print(f" {lang}: {count}") |
| | print("=" * 60) |
| |
|
| |
|
| | if __name__ == "__main__": |
| | main() |
| |
|