File size: 5,094 Bytes
4f3431c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0d95119
4f3431c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
import re
import json

from tqdm import tqdm
from loguru import logger

from pathlib import Path
from typing import Tuple, List


project_root = Path(__file__).parent.parent.parent

problem_tag = 'Problem'
solution_tag = 'Solution'


def clean_text(text: str) -> str:
    clean_texts = ["-SPB-", "SPB-", "-FIN-", "-POL-", "POL-", "EST-", "-DEN-", "DEN-", "NOR-", "-LAT-"]
    for clean_text in clean_texts:
        text = text.replace(clean_text, "")
    return text


def analyze(text: str) -> Tuple[List, int]:
    """
    Analyze the text and return the tags and problem number.
    Args:
        text (str): The markdown text to analyze.
    Returns:
        Tuple[List, int]: A tuple containing the tags and problem number.
    """
    problem_pattern = re.compile(r'(?:(?:\n|# )Problem\s+(\d+)(?:\.|\:)?|\n(\d+)\.|\n\#+\s+((?:A|C|G|N)\-\d+)\s+[A-Z]{3}\n)', re.IGNORECASE)
    solution_pattern = re.compile(r'(?:\n|# )Solution(?:\s+(\d+)\:?|\.|:|\n)', re.IGNORECASE)

    tags = []
    tags.extend([(x, problem_tag) for x in problem_pattern.finditer(text)])
    problem_num = len(tags)

    tags.extend([(x, solution_tag) for x in solution_pattern.finditer(text)])
    solution_num = len(tags) - problem_num

    tags.sort(key=lambda x: x[0].start())
    return tags, problem_num, solution_num


def segment(text: str, tags):
    starts = []
    ends = []

    for i in range(len(tags)):
        starts.append(tags[i][0].end())
        if i + 1 < len(tags):
            ends.append(tags[i + 1][0].start())
        else:
            ends.append(len(text))

    return [text[start:end].strip().strip('#').strip() for start, end in zip(starts, ends)]


def join(tags, segments, by_tag=False):
    problem, solution = '', ''
    problem_label, problem_match, solution_match = '', '', ''
    pairs = []

    if by_tag:
        seen = {}
        for tag, segment in zip(tags, segments):
            match = tag[0].group(0)
            if match in seen:
                problem, problem_match, problem_label = seen[match]
                pairs.append((problem, segment, problem_label, problem_match, match))
            else:
                seen[match] = (segment, match, tag[0].group(1) or tag[0].group(2) or tag[0].group(3))               
    else:
        for tag, segment in zip(tags, segments):
            if tag[1] == problem_tag:
                problem = segment
                problem_match = tag[0].group(0)
                problem_label = tag[0].group(1) or tag[0].group(2) or tag[0].group(3)
            else:
                solution = segment
                if solution.strip() == '':
                    continue

                solution_match = tag[0].group(0)
                pairs.append((problem, solution, problem_label, problem_match, solution_match))
            
    return pairs


def write_pairs(output_file: Path, pairs):
    year = re.search(r'(\d{2})', output_file.stem).group(1)
    year = f"20{year}" if int(year) >= 0 and int(year) <= 24 else f"19{year}"

    problem_type_mapping = {
        "A": "Algebra",
        "C": "Combinatorics",
        "G": "Geometry",
        "N": "Number Theory"
    }

    output_jsonl_text = ""
    for problem, solution, problem_label, problem_match, solution_match in pairs:
        output_jsonl_text += json.dumps(
            {
                'year': year,
                'tier': "T3",
                'problem_label': problem_label,
                'problem_type': problem_type_mapping.get(problem_label[0]),
                'exam': 'BalticWay',
                'problem': problem,
                'solution': solution,
                'metadata': {
                    'resource_path': output_file.relative_to(project_root).as_posix(),
                    'problem_match': problem_match,
                    'solution_match': solution_match
                }
            },
            ensure_ascii=False
        ) + '\n'

    output_file.write_text(output_jsonl_text, encoding="utf-8")


def main():
    compet_base_path = Path(__file__).resolve().parent.parent
    compet_md_path = compet_base_path / "md"
    seg_output_path = compet_base_path / "segmented"

    total_problem_count = 0
    total_solution_count = 0

    for bw_md in tqdm(list(compet_md_path.glob('**/*.md')), desc='Segmenting'):
        output_file = seg_output_path / bw_md.relative_to(compet_md_path).with_suffix('.jsonl')
        output_file.parent.mkdir(parents=True, exist_ok=True)

        text = '\n' + clean_text(bw_md.read_text(encoding="utf-8"))

        tags, problem_num, solution_num = analyze(text)

        segments = segment(text, tags)
        pairs = join(tags, segments, by_tag=solution_num == 0)     
        if pairs and problem_num > 0:
            write_pairs(output_file, pairs)

            total_problem_count += problem_num
            total_solution_count += len(pairs)
        else:
            logger.warning(f"No problem found in {bw_md}")
        
    logger.info(f"Total problem count: {total_problem_count}")
    logger.info(f"Total solution count: {total_solution_count}")


if __name__ == '__main__':
    main()