|
|
|
|
|
|
|
|
|
|
|
|
|
|
import json |
|
|
import re |
|
|
from pathlib import Path |
|
|
from typing import List, Tuple |
|
|
|
|
|
|
|
|
problem_tag = "Problem" |
|
|
solution_tag = "Solution" |
|
|
|
|
|
|
|
|
|
|
|
def clean_text(text: str): |
|
|
text = re.sub(r"Problem A1", "Problem 1", text) |
|
|
text = re.sub(r"Problem A2", "Problem 2", text) |
|
|
text = re.sub(r"Problem A3", "Problem 3", text) |
|
|
text = re.sub(r"Problem B1", "Problem 4", text) |
|
|
text = re.sub(r"Problem B2", "Problem 5", text) |
|
|
text = re.sub(r"Problem B3", "Problem 6", text) |
|
|
return text |
|
|
|
|
|
|
|
|
def segment_exams(text: str): |
|
|
matchs = list( |
|
|
|
|
|
re.finditer( |
|
|
r"^#+\s*.+USAMO\s*(\d{4})", |
|
|
text, |
|
|
flags=re.IGNORECASE | re.MULTILINE, |
|
|
) |
|
|
) |
|
|
|
|
|
exams = {} |
|
|
for i, m in enumerate(matchs): |
|
|
if int(m.group(1)) >= 1996: |
|
|
continue |
|
|
|
|
|
year = m.group(1) |
|
|
exam_text = text[ |
|
|
m.end() : matchs[i + 1].start() if i + 1 < len(matchs) else len(text) |
|
|
] |
|
|
|
|
|
exams[year] = exam_text.strip() |
|
|
|
|
|
return exams |
|
|
|
|
|
|
|
|
def analyze(text: str) -> Tuple[List, int]: |
|
|
""" |
|
|
Analyze the text and return the tags and problem number. |
|
|
|
|
|
Args: |
|
|
text (str): The markdown text to analyze. |
|
|
|
|
|
Returns: |
|
|
Tuple[List, int]: A tuple containing the tags and problem number. |
|
|
""" |
|
|
problem_pattern = re.compile(r"(?:\n|# )Problem\s+(\d+)", re.IGNORECASE) |
|
|
solution_pattern = re.compile(r"(?:\n|# )Solution", re.IGNORECASE) |
|
|
|
|
|
|
|
|
tags = [] |
|
|
tags.extend([(x, problem_tag) for x in problem_pattern.finditer(text)]) |
|
|
problem_num = len(tags) |
|
|
|
|
|
tags.extend([(x, solution_tag) for x in solution_pattern.finditer(text)]) |
|
|
|
|
|
|
|
|
tags.sort(key=lambda x: x[0].start()) |
|
|
return tags, problem_num |
|
|
|
|
|
|
|
|
def segment(text: str, tags): |
|
|
starts = [] |
|
|
ends = [] |
|
|
|
|
|
for i, (m, tag) in enumerate(tags): |
|
|
starts.append(tags[i][0].end()) |
|
|
if i + 1 < len(tags): |
|
|
ends.append(tags[i + 1][0].start()) |
|
|
else: |
|
|
ends.append(len(text)) |
|
|
|
|
|
return [ |
|
|
text[start:end].strip().strip("#").strip() for start, end in zip(starts, ends) |
|
|
] |
|
|
|
|
|
|
|
|
def join(tags, segments): |
|
|
problem, solution = "", "" |
|
|
problem_label, problem_match, solution_match = "", "", "" |
|
|
pairs = [] |
|
|
|
|
|
tag_classes = [_[1] for _ in tags] |
|
|
|
|
|
for (m, tag), (i, segment) in zip(tags, enumerate(segments)): |
|
|
if tag == problem_tag: |
|
|
problem = segment |
|
|
problem_match = m.group(0) |
|
|
problem_label = m.group(1) |
|
|
|
|
|
|
|
|
next_problem_index = 0 |
|
|
try: |
|
|
if problem_tag in tag_classes[i + 1 :]: |
|
|
next_problem_index = tag_classes.index(problem_tag, i + 1) |
|
|
else: |
|
|
next_problem_index = len(segments) |
|
|
except ValueError: |
|
|
next_problem_index = len(segments) |
|
|
|
|
|
if tag_classes[i + 1 : next_problem_index].count(solution_tag) == 0: |
|
|
solution = "" |
|
|
solution_match = "" |
|
|
pairs.append( |
|
|
(problem, solution, problem_label, problem_match, solution_match) |
|
|
) |
|
|
else: |
|
|
solution = segment |
|
|
solution_match = m.group(0) |
|
|
pairs.append( |
|
|
(problem, solution, problem_label, problem_match, solution_match) |
|
|
) |
|
|
|
|
|
return pairs |
|
|
|
|
|
|
|
|
def write_pairs(project_root: Path, output_file: Path, pairs): |
|
|
output_jsonl_text = "" |
|
|
for year, problems in pairs: |
|
|
for ( |
|
|
problem, |
|
|
solution, |
|
|
problem_label, |
|
|
problem_match, |
|
|
solution_match, |
|
|
) in problems: |
|
|
output_jsonl_text += ( |
|
|
json.dumps( |
|
|
{ |
|
|
"year": year, |
|
|
"tier": "T1", |
|
|
"problem_label": problem_label, |
|
|
"problem_type": None, |
|
|
"exam": "USAMO", |
|
|
"problem": problem, |
|
|
"solution": solution, |
|
|
"metadata": { |
|
|
"resource_path": output_file.relative_to( |
|
|
project_root |
|
|
).as_posix(), |
|
|
"problem_match": problem_match, |
|
|
"solution_match": solution_match, |
|
|
}, |
|
|
}, |
|
|
ensure_ascii=False, |
|
|
) |
|
|
+ "\n" |
|
|
) |
|
|
|
|
|
output_file.write_text(output_jsonl_text, encoding="utf-8") |
|
|
|
|
|
|
|
|
if __name__ == "__main__": |
|
|
compet_base_path = Path(__file__).resolve().parent.parent |
|
|
compet_md_path = compet_base_path / "md" |
|
|
seg_output_path = compet_base_path / "segmented" |
|
|
project_root = compet_base_path.parent |
|
|
|
|
|
for md_file in list(compet_md_path.glob("**/en-USAMO-1972-2003.md")): |
|
|
output_file = seg_output_path / md_file.relative_to(compet_md_path).with_suffix( |
|
|
".jsonl" |
|
|
) |
|
|
output_file.parent.mkdir(parents=True, exist_ok=True) |
|
|
|
|
|
|
|
|
markdown_text = md_file.read_text(encoding="utf-8") |
|
|
markdown_text = clean_text(markdown_text) |
|
|
|
|
|
|
|
|
pairs = [] |
|
|
exams = segment_exams(markdown_text) |
|
|
for year, exam_text in exams.items(): |
|
|
tags, problem_num = analyze(exam_text) |
|
|
segments = segment(exam_text, tags) |
|
|
inner_pairs = join(tags, segments) |
|
|
pairs.append((year, inner_pairs)) |
|
|
|
|
|
write_pairs(project_root, output_file, pairs) |
|
|
|