File size: 3,511 Bytes
24a410c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
import os
import re

import pandas as pd
from bs4 import BeautifulSoup


def extract_puzzle_data(html):
    """extract data from HTML content"""
    soup = BeautifulSoup(html, 'html.parser')

    # Extract task description
    task_description_section = soup.find('div', id='tabs-2') or ""
    task_description = task_description_section.get_text(strip=True).replace('\xa0', ' ')
    task_description = task_description.removeprefix("Backstory and Goal")
    task_description = task_description.partition("Remember, as with all")[0]

    categories = [cc.get_text(strip=True).replace('\xa0', ' ') for cc in soup.find_all('td', class_='answergrid_head')]

    # Extract clues, and clean them
    clues = []
    for clue_div in soup.find_all('div', class_='clue'):
        clue_raw = clue_div.get_text(strip=True).replace('\xa0', ' ')
        # Remove numbers at the beginning of the string followed by a period and whitespace
        cleaned_clue = re.sub(r'^\d+\.\s*', '', clue_raw)
        clues.append(cleaned_clue)

    # Extract label names such as from labelboxh
    label_categories = dict(label_a=[])
    for label in soup.find_all('td', class_='labelboxh'):
        if label['id'].startswith("labelleftA"):
            label_categories["label_a"].append(label.get_text(strip=True).replace('\xa0', ' '))
    for letter in "bcd":
        pattern = re.compile(f'label{letter}_ary' + r'\[\d+]\s*=\s*"([^"]+)";')
        items = pattern.findall(html)
        label_categories[f"label_{letter}"] = items
    return dict(story=task_description,
                clues=clues,
                categories=categories,
                **label_categories)


global_stories = set()
global_clues = set()


def process_one(difficulty, grid_size):
    puzzle_data = []
    with open(f'urls/{difficulty}{grid_size}.txt') as rr:
        all_paths = [p.strip() for p in rr]
    dir_path = f'htmls/{difficulty}{grid_size}/'
    for c, puzzle_url in enumerate(all_paths):
        filename = puzzle_url.removeprefix("https://logic.puzzlebaron.com/")
        if c % 200 == 0:
            print(f"{c=}")
        file_path = os.path.join(dir_path, filename)
        with open(file_path, 'r', encoding='utf-8') as file:
            html_content = file.read()
            data = extract_puzzle_data(html_content)
            if len(global_clues.intersection(data['clues'])) >= 3:
                continue
            # elif len(global_clues.intersection(data['clues'])) >= 4:
            #     print("FAIL:", difficulty, grid_size)
            #     print(filename)
            #     print(global_clues.intersection(data['clues']))
            #     continue
            #     raise RuntimeError(global_clues.intersection(data['clues']))
            global_clues.update(data['clues'])
            data['grid_size'] = grid_size
            data['difficulty'] = difficulty
            data['url'] = puzzle_url
            puzzle_data.append(data)
    return puzzle_data


OUTPUT_DIR = "dataframes"


def main():
    if not os.path.exists(OUTPUT_DIR):
        os.makedirs(OUTPUT_DIR)
    for grid_size in ['4x7', '4x6', '4x5', '4x4', '3x5', '3x4']:
        for difficulty in ['challenging', 'moderate', 'easy']:
            puzzle_data = process_one(difficulty, grid_size)
            df = pd.DataFrame(puzzle_data)
            jsonl_file_path = f'{OUTPUT_DIR}/{difficulty}{grid_size}.jsonl'
            df.to_json(jsonl_file_path, orient='records', lines=True)
            print(f'Data saved to {jsonl_file_path}', df.shape)


main()