File size: 5,203 Bytes
e76559c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3f586b5
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
"""Main orchestration module for Finesse Benchmark Database Generator



This script serves as the 'Conductor' that orchestrates the entire pipeline: from configuration setup

through atomic bead generation to final JSONL serialization. It ties together config, chunker, and writer

for end-to-end execution of our 'Gemstone Necklace' production process.



Key Principles:

- Ensures all modules use the shared CONFIG for consistency.

- Executes chunking first (memory-intensive), then writing (I/O-focused).

- Comprehensive logging for audit trail and debugging.

- Designed for one-shot full generation; scale via external orchestration if needed.



Usage:

python main.py

# Runs the full pipeline: generates ~100,000 strings of beads and writes to probes_atomic.jsonl

"""

import logging

from .config import ProbeConfig
from .chunker import generate_all_strings_of_beads
from .writer import write_strings_to_probes_atomic

# Configure logging for the entire pipeline
logging.basicConfig(level=logging.INFO, format='%(asctime)s - %(levelname)s - %(message)s')
logger = logging.getLogger(__name__)

def generate_dataset(config: ProbeConfig):
    logger.info(f"Config set: {len(config.languages)} languages, {config.chunk_token_size}-token beads, "
                f"{config.samples_per_language} strings per language")
    
    # Step 2: Generate all structured 'strings of beads' using the chunker
    logger.info("Starting bead generation with chunker...")
    all_strings_of_beads = generate_all_strings_of_beads(config)
    total_strings = len(all_strings_of_beads)
    logger.info(f"Bead generation complete: {total_strings} total strings produced "
                f"(~{total_strings / len(config.languages):.0f} per language)")
    
    # Step 3: Serialize to the atomic probes JSONL file using the writer
    logger.info(f"Starting serialization to {config.output_file}...")
    write_strings_to_probes_atomic(config, all_strings_of_beads)
    logger.info(f"Serialization complete: All {total_strings} strings saved with metadata.")
    
    # Final summary
    estimated_beads = sum(len(s['beads']) for s in all_strings_of_beads)
    logger.info("=== Pipeline Summary (Test Run) ===")
    logger.info(f"- Total strings: {total_strings}")
    logger.info(f"- Total atomic beads: {estimated_beads}")
    logger.info(f"- Coverage: {len(set(s['source']['lang'] for s in all_strings_of_beads))} languages")
    logger.info(f"- Output file: {config.output_file}")
    logger.info("For full production: Increase samples_per_language and add more languages in config.")
    logger.info("=== Finesse Benchmark Database Generation Example Finished Successfully ===")


def main() -> None:
    """Execute the full Finesse benchmark database generation pipeline as a library example.

    

    This demonstrates how users would create a ProbeConfig, generate data, and write output.

    For production, adjust config parameters as needed (e.g., more languages, larger samples).

    """
    logger.info("=== Finesse Benchmark Database Pipeline Started (Library Example) ===")
    
    # Step 1: User creates and configures ProbeConfig for this run
    logger.info("Creating test configuration...")
    test_config = ProbeConfig(
        languages=['en'],  # Single language for quick demo
        samples_per_language=5,  # Small sample for testing
        chunk_token_size=64,
        tokenizer_name="google-bert/bert-base-multilingual-cased",
        output_file="test_probes_atomic.jsonl",
        seed=42
    )
    logger.info(f"Config set: {len(test_config.languages)} languages, {test_config.chunk_token_size}-token beads, "
                f"{test_config.samples_per_language} strings per language")
    
    # Step 2: Generate all structured 'strings of beads' using the chunker
    logger.info("Starting bead generation with chunker...")
    all_strings_of_beads = generate_all_strings_of_beads(test_config)
    total_strings = len(all_strings_of_beads)
    logger.info(f"Bead generation complete: {total_strings} total strings produced "
                f"(~{total_strings / len(test_config.languages):.0f} per language)")
    
    # Step 3: Serialize to the atomic probes JSONL file using the writer
    logger.info(f"Starting serialization to {test_config.output_file}...")
    write_strings_to_probes_atomic(test_config, all_strings_of_beads)
    logger.info(f"Serialization complete: All {total_strings} strings saved with metadata.")
    
    # Final summary
    estimated_beads = sum(len(s['beads']) for s in all_strings_of_beads)
    logger.info("=== Pipeline Summary (Test Run) ===")
    logger.info(f"- Total strings: {total_strings}")
    logger.info(f"- Total atomic beads: {estimated_beads}")
    logger.info(f"- Coverage: {len(set(s['source']['lang'] for s in all_strings_of_beads))} languages")
    logger.info(f"- Output file: {test_config.output_file}")
    logger.info("For full production: Increase samples_per_language and add more languages in config.")
    logger.info("=== Finesse Benchmark Database Generation Example Finished Successfully ===")

if __name__ == "__main__":
    main()