enzoescipy's picture
Upload folder using huggingface_hub
e76559c verified
"""Main orchestration module for Finesse Benchmark Database Generator
This script serves as the 'Conductor' that orchestrates the entire pipeline: from configuration setup
through atomic bead generation to final JSONL serialization. It ties together config, chunker, and writer
for end-to-end execution of our 'Gemstone Necklace' production process.
Key Principles:
- Ensures all modules use the shared CONFIG for consistency.
- Executes chunking first (memory-intensive), then writing (I/O-focused).
- Comprehensive logging for audit trail and debugging.
- Designed for one-shot full generation; scale via external orchestration if needed.
Usage:
python main.py
# Runs the full pipeline: generates ~100,000 strings of beads and writes to probes_atomic.jsonl
"""
import logging
from .config import ProbeConfig
from .chunker import generate_all_strings_of_beads
from .writer import write_strings_to_probes_atomic
# Configure logging for the entire pipeline
logging.basicConfig(level=logging.INFO, format='%(asctime)s - %(levelname)s - %(message)s')
logger = logging.getLogger(__name__)
def generate_dataset(config: ProbeConfig):
logger.info(f"Config set: {len(config.languages)} languages, {config.chunk_token_size}-token beads, "
f"{config.samples_per_language} strings per language")
# Step 2: Generate all structured 'strings of beads' using the chunker
logger.info("Starting bead generation with chunker...")
all_strings_of_beads = generate_all_strings_of_beads(config)
total_strings = len(all_strings_of_beads)
logger.info(f"Bead generation complete: {total_strings} total strings produced "
f"(~{total_strings / len(config.languages):.0f} per language)")
# Step 3: Serialize to the atomic probes JSONL file using the writer
logger.info(f"Starting serialization to {config.output_file}...")
write_strings_to_probes_atomic(config, all_strings_of_beads)
logger.info(f"Serialization complete: All {total_strings} strings saved with metadata.")
# Final summary
estimated_beads = sum(len(s['beads']) for s in all_strings_of_beads)
logger.info("=== Pipeline Summary (Test Run) ===")
logger.info(f"- Total strings: {total_strings}")
logger.info(f"- Total atomic beads: {estimated_beads}")
logger.info(f"- Coverage: {len(set(s['source']['lang'] for s in all_strings_of_beads))} languages")
logger.info(f"- Output file: {config.output_file}")
logger.info("For full production: Increase samples_per_language and add more languages in config.")
logger.info("=== Finesse Benchmark Database Generation Example Finished Successfully ===")
def main() -> None:
"""Execute the full Finesse benchmark database generation pipeline as a library example.
This demonstrates how users would create a ProbeConfig, generate data, and write output.
For production, adjust config parameters as needed (e.g., more languages, larger samples).
"""
logger.info("=== Finesse Benchmark Database Pipeline Started (Library Example) ===")
# Step 1: User creates and configures ProbeConfig for this run
logger.info("Creating test configuration...")
test_config = ProbeConfig(
languages=['en'], # Single language for quick demo
samples_per_language=5, # Small sample for testing
chunk_token_size=64,
tokenizer_name="google-bert/bert-base-multilingual-cased",
output_file="test_probes_atomic.jsonl",
seed=42
)
logger.info(f"Config set: {len(test_config.languages)} languages, {test_config.chunk_token_size}-token beads, "
f"{test_config.samples_per_language} strings per language")
# Step 2: Generate all structured 'strings of beads' using the chunker
logger.info("Starting bead generation with chunker...")
all_strings_of_beads = generate_all_strings_of_beads(test_config)
total_strings = len(all_strings_of_beads)
logger.info(f"Bead generation complete: {total_strings} total strings produced "
f"(~{total_strings / len(test_config.languages):.0f} per language)")
# Step 3: Serialize to the atomic probes JSONL file using the writer
logger.info(f"Starting serialization to {test_config.output_file}...")
write_strings_to_probes_atomic(test_config, all_strings_of_beads)
logger.info(f"Serialization complete: All {total_strings} strings saved with metadata.")
# Final summary
estimated_beads = sum(len(s['beads']) for s in all_strings_of_beads)
logger.info("=== Pipeline Summary (Test Run) ===")
logger.info(f"- Total strings: {total_strings}")
logger.info(f"- Total atomic beads: {estimated_beads}")
logger.info(f"- Coverage: {len(set(s['source']['lang'] for s in all_strings_of_beads))} languages")
logger.info(f"- Output file: {test_config.output_file}")
logger.info("For full production: Increase samples_per_language and add more languages in config.")
logger.info("=== Finesse Benchmark Database Generation Example Finished Successfully ===")
if __name__ == "__main__":
main()