| import re |
| import json |
|
|
| from tqdm import tqdm |
| from loguru import logger |
|
|
| from pathlib import Path |
| from typing import Tuple, List |
|
|
|
|
| project_root = Path(__file__).parent.parent.parent |
|
|
| problem_tag = "Problem" |
| solution_tag = "Solution" |
|
|
|
|
| def clean_text(text: str) -> str: |
| text = text.replace("\n\n## Canadian Mathematical Olympiad 2021", "") |
| text = text.replace( |
| "1. no colour is assigned to two regions that share an edge;\n2. for each $i", |
| "1). no colour is assigned to two regions that share an edge;\n2). for each $i", |
| ) |
| text = re.sub(r"^Comment(?:\s+\d+)?\..+?\n$", "", text, flags=re.IGNORECASE | re.MULTILINE) |
| return text |
|
|
|
|
| def find_problem_with_solution( |
| text: str, problem_parttern: re.Pattern, solution_pattern: re.Pattern |
| ) -> int: |
| """ |
| Find the problem with solution start position in the text. |
| Args: |
| text (str): The text to search. |
| Returns: |
| int: The start position of the problem with solution. |
| """ |
| matchs = list(problem_parttern.finditer(text)) |
|
|
| for index, match in enumerate(matchs): |
| section_end_position = ( |
| matchs[index + 1].start() if index + 1 < len(matchs) else len(text) |
| ) |
| if solution_pattern.search(text[match.start() : section_end_position]): |
| return match.start() |
|
|
|
|
| def analyze(text: str) -> Tuple[List, int]: |
| """ |
| Analyze the text and return the tags and problem number. |
| Args: |
| text (str): The markdown text to analyze. |
| Returns: |
| Tuple[List, int]: A tuple containing the tags and problem number. |
| """ |
| problem_pattern = re.compile( |
| r"(?:\n|\n#+ )(?:Problem\s*(\d+)|Problem No\.\s*(\d+)|P(\d+)|(\d+))(?:\.|\:)", |
| re.IGNORECASE, |
| ) |
| solution_pattern = re.compile( |
| r"(?:\n|\n#+ )(?:Solution|First Solution|Second Solution|Alternate Solution)\s*\d*(?:\.|\:)?", |
| re.IGNORECASE, |
| ) |
|
|
| start_position = ( |
| find_problem_with_solution(text, problem_pattern, solution_pattern) or 0 |
| ) |
|
|
| tags = [] |
| tags.extend( |
| [(x, problem_tag) for x in problem_pattern.finditer(text, start_position)] |
| ) |
| problem_num = len(tags) |
|
|
| tags.extend( |
| [(x, solution_tag) for x in solution_pattern.finditer(text, start_position)] |
| ) |
| tags.sort(key=lambda x: x[0].start()) |
| return tags, problem_num |
|
|
|
|
| def segment(text: str, tags): |
| starts = [] |
| ends = [] |
|
|
| for i in range(len(tags)): |
| starts.append(tags[i][0].end()) |
| if i + 1 < len(tags): |
| ends.append(tags[i + 1][0].start()) |
| else: |
| ends.append(len(text)) |
|
|
| return [ |
| text[start:end].strip().strip("#").strip() for start, end in zip(starts, ends) |
| ] |
|
|
|
|
| def join(tags, segments): |
| problem, solution = "", "" |
| problem_label, problem_match, solution_match = "", "", "" |
| pairs = [] |
|
|
| for tag, segment in zip(tags, segments): |
| if tag[1] == problem_tag: |
| problem = segment |
| problem_match = tag[0].group(0) |
| problem_label = ( |
| tag[0].group(1) or tag[0].group(2) or tag[0].group(3) or tag[0].group(4) |
| ) |
| else: |
| solution = segment |
| solution_match = tag[0].group(0) |
| pairs.append( |
| (problem, solution, problem_label, problem_match, solution_match) |
| ) |
|
|
| return pairs |
|
|
|
|
| def write_pairs(output_file: Path, pairs): |
| year = re.search(r"(\d{4})", output_file.stem).group(1) |
|
|
| output_jsonl_text = "" |
| for problem, solution, problem_label, problem_match, solution_match in pairs: |
| output_jsonl_text += ( |
| json.dumps( |
| { |
| "year": year, |
| "tier": "T2", |
| "problem_label": problem_label, |
| "problem_type": None, |
| "exam": "Canada_MO", |
| "problem": problem, |
| "solution": solution, |
| "metadata": { |
| "resource_path": output_file.relative_to( |
| project_root |
| ).as_posix(), |
| "problem_match": problem_match, |
| "solution_match": solution_match, |
| }, |
| }, |
| ensure_ascii=False, |
| ) |
| + "\n" |
| ) |
|
|
| output_file.write_text(output_jsonl_text, encoding="utf-8") |
|
|
|
|
| def main(): |
| compet_base_path = Path(__file__).resolve().parent.parent |
| compet_md_path = compet_base_path / "md" |
| seg_output_path = compet_base_path / "segmented" |
|
|
| total_problem_count = 0 |
| total_solution_count = 0 |
|
|
| for cmo_md in tqdm(list(compet_md_path.glob("**/*.md")), desc="Segmenting"): |
| year = re.search(r"(\d{4})", cmo_md.stem).group(1) |
| |
| if int(year) not in list(range(2002, 2007)) + list(range(2008, 2025 + 1)): |
| continue |
|
|
| output_file = seg_output_path / cmo_md.relative_to(compet_md_path).with_suffix( |
| ".jsonl" |
| ) |
| output_file.parent.mkdir(parents=True, exist_ok=True) |
|
|
| text = "\n" + clean_text(cmo_md.read_text(encoding="utf-8")) |
|
|
| tags, problem_num = analyze(text) |
|
|
| segments = segment(text, tags) |
| pairs = join(tags, segments) |
| if pairs and problem_num > 0: |
| write_pairs(output_file, pairs) |
|
|
| total_problem_count += problem_num |
| total_solution_count += len(pairs) |
| else: |
| logger.warning(f"No problem found in {cmo_md}") |
|
|
| logger.info(f"Total problem count: {total_problem_count}") |
| logger.info(f"Total solution count: {total_solution_count}") |
|
|
|
|
| if __name__ == "__main__": |
| main() |
|
|