data / gepa_app.py
infinex's picture
Uploading dataset files from the local data folder.
b48b2f3 verified
import streamlit as st
import dspy
import pandas as pd
import numpy as np
import re
import httpx
import json
from openai import OpenAI
from st_aggrid import AgGrid, GridOptionsBuilder, GridUpdateMode, DataReturnMode
from typing import Optional, Dict, Any, List
import os
# --- Page Configuration ---
st. set_page_config(
layout="wide",
page_title="GEPA Regex Optimizer",
page_icon="๐Ÿงฌ",
initial_sidebar_state="expanded"
)
# --- Session State Initialization ---
DEFAULT_STATE = {
'dataset': None,
'selected_indices': [], # Track selected row indices for training
'optimized_program': None,
'optimization_history': [],
'config': {
'model_name': 'gpt-4o',
'api_key': '',
'base_url': 'https://api.openai.com/v1',
'timeout': 30,
'max_retries': 3,
'temperature': 0.7,
'max_tokens': 1024,
},
'gepa_config': {
'num_iterations': 5,
'num_candidates': 3,
'early_stopping_threshold': 0.95,
},
'prompts': {
'system_instruction': "You are a Regex Expert. Given the input text, provide a high-precision Python regex pattern to extract the target text.",
'gepa_meta_prompt': "Focus on precision. If the feedback says the match was too broad, use more specific character classes or anchors. If it missed the target, suggest more flexible patterns.",
'output_description': "A Python-compatible regular expression",
},
'train_test_split': 0.8,
'regex_flags': [],
}
for key, value in DEFAULT_STATE.items():
if key not in st.session_state:
st.session_state[key] = value
# --- Configuration Manager ---
class ConfigManager:
"""Manages application configuration with persistence."""
CONFIG_FILE = "gepa_config.json"
@staticmethod
def save_config():
"""Save current configuration to file."""
config_data = {
'config': st.session_state. config,
'gepa_config': st.session_state. gepa_config,
'prompts': st.session_state.prompts,
'train_test_split': st.session_state. train_test_split,
'regex_flags': st. session_state.regex_flags,
}
try:
with open(ConfigManager.CONFIG_FILE, 'w') as f:
json.dump(config_data, f, indent=2)
return True
except Exception as e:
st.error(f"Failed to save config: {e}")
return False
@staticmethod
def load_config():
"""Load configuration from file."""
try:
if os.path.exists(ConfigManager.CONFIG_FILE):
with open(ConfigManager.CONFIG_FILE, 'r') as f:
config_data = json.load(f)
for key, value in config_data. items():
if key in st.session_state:
if isinstance(value, dict):
st. session_state[key].update(value)
else:
st. session_state[key] = value
return True
except Exception as e:
st.warning(f"Failed to load config: {e}")
return False
@staticmethod
def reset_to_defaults():
"""Reset all configuration to defaults."""
for key, value in DEFAULT_STATE.items():
if key not in ['dataset', 'optimized_program', 'optimization_history']:
st.session_state[key] = value. copy() if isinstance(value, (dict, list)) else value
# --- LLM Setup ---
def setup_dspy() -> bool:
"""Configure DSPy with current settings."""
config = st.session_state. config
try:
http_client = httpx.Client(
timeout=config['timeout'],
limits=httpx.Limits(max_retries=config['max_retries'])
)
custom_openai_client = OpenAI(
api_key=config['api_key'] or os.getenv("OPENAI_API_KEY", "empty"),
base_url=config['base_url'] or None,
http_client=http_client
)
lm = dspy.LM(
model=config['model_name'],
client=custom_openai_client,
temperature=config['temperature'],
max_tokens=config['max_tokens']
)
dspy.configure(lm=lm)
return True
except Exception as e:
st. error(f"LLM Configuration Error: {e}")
return False
# --- Metric Function ---
def create_regex_metric(flags: list):
"""Factory function to create metric with configurable regex flags."""
compiled_flags = 0
for flag in flags:
compiled_flags |= getattr(re, flag, 0)
def regex_metric_with_feedback(example, prediction, trace=None):
"""GEPA Metric with rich feedback for regex optimization."""
target = example. ground_truth. strip()
raw_text = example. raw_text
pred_pattern = getattr(prediction, 'regex_pattern', '').strip()
# Handle missing output
if not pred_pattern:
feedback = (
f"No regex pattern provided. Target text: '{target}'. "
"Please output a valid Python regex string."
)
return dspy. Prediction(score=0.0, feedback=feedback)
# Syntax validation
try:
compiled = re.compile(pred_pattern, compiled_flags)
except re.error as e:
feedback = (
f"Invalid regex: '{pred_pattern}'. "
f"Error: {str(e)}. Check syntax and escape characters."
)
return dspy. Prediction(score=0.0, feedback=feedback)
# Match evaluation
match = compiled.search(raw_text)
extracted = match.group(0) if match else ""
if extracted == target:
return dspy.Prediction(
score=1.0,
feedback=f"Perfect match! Correctly extracted '{target}'."
)
# Failure analysis
score = 0.0
feedback = f"Pattern '{pred_pattern}' produced incorrect result.\n"
if not match:
feedback += f"NO MATCH found. Target: '{target}'."
elif target in extracted:
score = 0.3
feedback += (
f"TOO BROAD: Extracted '{extracted}' contains target '{target}' "
"plus extra characters. Use stricter boundaries or non-greedy quantifiers."
)
elif extracted in target:
score = 0.3
feedback += (
f"TOO NARROW: Extracted '{extracted}' but target is '{target}'. "
"Make pattern more inclusive."
)
else:
feedback += f"WRONG MATCH: Got '{extracted}' instead of '{target}'."
feedback += "\nAnalyze the target structure to isolate it uniquely."
return dspy.Prediction(score=score, feedback=feedback)
return regex_metric_with_feedback
# --- DSPy Program ---
class RegexSignature(dspy. Signature):
"""Dynamic signature for regex generation."""
raw_text = dspy. InputField()
regex_pattern = dspy.OutputField()
class RegexGenerator(dspy.Module):
"""Configurable regex generation module."""
def __init__(self, doc: str, output_desc: str):
super().__init__()
self.predictor = dspy.Predict(RegexSignature)
self.predictor.signature.__doc__ = doc
self.predictor.signature.regex_pattern. desc = output_desc
def forward(self, raw_text: str):
return self. predictor(raw_text=raw_text)
# --- Sidebar Configuration ---
def render_sidebar():
"""Render the configuration sidebar."""
with st.sidebar:
st.title("โš™๏ธ Configuration")
# Config management buttons
col1, col2, col3 = st.columns(3)
with col1:
if st.button("๐Ÿ’พ Save", use_container_width=True):
if ConfigManager.save_config():
st.success("Saved!")
with col2:
if st.button("๐Ÿ“‚ Load", use_container_width=True):
if ConfigManager.load_config():
st.success("Loaded!")
st.rerun()
with col3:
if st.button("๐Ÿ”„ Reset", use_container_width=True):
ConfigManager.reset_to_defaults()
st.rerun()
st.divider()
# LLM Configuration
with st.expander("๐Ÿค– LLM Settings", expanded=True):
st.session_state.config['model_name'] = st.text_input(
"Model Name",
value=st.session_state.config['model_name'],
help="e.g., gpt-4o, gpt-3.5-turbo, claude-3-opus"
)
st.session_state.config['api_key'] = st.text_input(
"API Key",
value=st.session_state.config['api_key'],
type="password",
help="Leave empty to use OPENAI_API_KEY env var"
)
st.session_state.config['base_url'] = st.text_input(
"Base URL",
value=st.session_state.config['base_url'],
help="Custom API endpoint (e.g., for Azure, local models)"
)
col1, col2 = st.columns(2)
with col1:
st.session_state.config['timeout'] = st.number_input(
"Timeout (s)",
min_value=5,
max_value=300,
value=st.session_state.config['timeout']
)
with col2:
st.session_state.config['max_retries'] = st.number_input(
"Max Retries",
min_value=0,
max_value=10,
value=st.session_state.config['max_retries']
)
col1, col2 = st.columns(2)
with col1:
st.session_state.config['temperature'] = st.slider(
"Temperature",
min_value=0.0,
max_value=2.0,
value=st. session_state.config['temperature'],
step=0.1
)
with col2:
st.session_state.config['max_tokens'] = st.number_input(
"Max Tokens",
min_value=64,
max_value=8192,
value=st.session_state.config['max_tokens']
)
# GEPA Optimizer Settings
with st. expander("๐Ÿงฌ GEPA Optimizer", expanded=False):
st.session_state.gepa_config['num_iterations'] = st.slider(
"Iterations",
min_value=1,
max_value=20,
value=st. session_state.gepa_config['num_iterations'],
help="Number of optimization iterations"
)
st.session_state. gepa_config['num_candidates'] = st.slider(
"Candidates per Iteration",
min_value=1,
max_value=10,
value=st.session_state.gepa_config['num_candidates'],
help="Number of candidate patterns to evaluate"
)
st. session_state.gepa_config['early_stopping_threshold'] = st.slider(
"Early Stopping Threshold",
min_value=0.5,
max_value=1.0,
value=st.session_state.gepa_config['early_stopping_threshold'],
step=0.05,
help="Stop if this score is reached"
)
# Prompt Configuration
with st.expander("๐Ÿ“ Prompts", expanded=False):
st.session_state.prompts['system_instruction'] = st.text_area(
"System Instruction",
value=st.session_state.prompts['system_instruction'],
height=100,
help="Initial instruction for regex generation"
)
st.session_state. prompts['gepa_meta_prompt'] = st.text_area(
"GEPA Evolution Prompt",
value=st.session_state.prompts['gepa_meta_prompt'],
height=100,
help="Instructions for GEPA's prompt evolution"
)
st.session_state. prompts['output_description'] = st. text_input(
"Output Field Description",
value=st.session_state.prompts['output_description'],
help="Description for the regex output field"
)
# Regex Configuration
with st. expander("๐Ÿ”ง Regex Options", expanded=False):
flag_options = ['IGNORECASE', 'MULTILINE', 'DOTALL', 'VERBOSE', 'ASCII']
st.session_state. regex_flags = st.multiselect(
"Regex Flags",
options=flag_options,
default=st.session_state. regex_flags,
help="Python regex flags to apply"
)
# Data Split Configuration
with st.expander("๐Ÿ“Š Data Settings", expanded=False):
st.session_state.train_test_split = st.slider(
"Train/Validation Split",
min_value=0.5,
max_value=0.95,
value=st.session_state.train_test_split,
step=0.05,
help="Proportion of data for training"
)
# --- Stratified Sampling Utility ---
def stratified_train_val_split(
df: pd.DataFrame,
train_ratio: float = 0.8,
stratify_column: str = 'ground_truth',
random_state: int = 42
) -> tuple:
"""
Perform stratified train/validation split.
Groups samples by ground_truth pattern and splits proportionally.
"""
np.random.seed(random_state)
# Create stratification groups based on ground_truth patterns
# Use first 50 chars of ground_truth as group key to handle similar patterns
df = df.copy()
df['_strat_key'] = df[stratify_column].apply(
lambda x: str(x)[:50] if pd.notna(x) and x != '' else '_empty_'
)
train_indices = []
val_indices = []
# Split each stratum
for _, group in df.groupby('_strat_key'):
indices = group.index.tolist()
np.random.shuffle(indices)
split_idx = max(1, int(len(indices) * train_ratio))
# Ensure at least one sample in validation if group has multiple samples
if len(indices) > 1 and split_idx == len(indices):
split_idx = len(indices) - 1
train_indices.extend(indices[:split_idx])
val_indices.extend(indices[split_idx:])
train_df = df.loc[train_indices].drop(columns=['_strat_key'])
val_df = df.loc[val_indices].drop(columns=['_strat_key']) if val_indices else pd.DataFrame()
return train_df, val_df
# --- Data Persistence ---
def save_annotated_data(df: pd.DataFrame, selected_indices: List[int], filepath: str) -> bool:
"""Save annotated data with selection state."""
try:
# Add selection column
save_df = df.copy()
save_df['_selected'] = save_df.index.isin(selected_indices)
if filepath.endswith('.json'):
save_df.to_json(filepath, orient='records', indent=2)
else:
save_df.to_csv(filepath, index=False)
return True
except Exception as e:
st.error(f"Failed to save data: {e}")
return False
def load_annotated_data(filepath: str) -> tuple:
"""Load annotated data with selection state."""
try:
df = pd.read_csv(filepath)
# Extract selection state if present
selected_indices = []
if '_selected' in df.columns:
selected_indices = df[df['_selected'] == True].index.tolist()
df = df.drop(columns=['_selected'])
# Ensure required columns
if 'text' not in df.columns:
raise ValueError("Dataset must have a 'text' column.")
if 'ground_truth' not in df.columns:
df['ground_truth'] = ''
return df, selected_indices
except Exception as e:
st.error(f"Failed to load data: {e}")
return None, []
# --- Main Application Tabs ---
def render_data_ingestion_tab():
"""Render the data ingestion tab."""
st.header("๐Ÿ“ฅ Data Ingestion & Annotation")
# File upload section
col1, col2 = st.columns([2, 1])
with col1:
uploaded = st.file_uploader(
"Upload Dataset",
type=["csv", "json", "xlsx"],
help="CSV/JSON/Excel with 'text' column (ground_truth optional, _selected for pre-selected rows)"
)
with col2:
st.markdown("**Expected Format:**")
st.code("text,ground_truth,_selected\n'Sample text','expected',true", language="csv")
if uploaded:
# Load based on file type
try:
df, selected_indices = load_annotated_data(uploaded)
if df is not None:
st.session_state.dataset = df.reset_index(drop=True)
st.session_state.selected_indices = selected_indices
st.success(f"โœ… Loaded {len(df)} samples ({len(selected_indices)} pre-selected)")
except Exception as e:
st.error(f"Failed to load file: {e}")
return
if st.session_state.dataset is not None:
df = st.session_state.dataset.copy()
st.subheader("๐Ÿ“ Annotate Ground Truth")
st.caption("Edit 'ground_truth' column and select rows (checkbox) to include in training/validation.")
# Prepare pre-selected rows for AgGrid
pre_selected_rows = st.session_state.get('selected_indices', [])
# Configure AgGrid
gb = GridOptionsBuilder.from_dataframe(df)
gb.configure_default_column(
resizable=True,
filterable=True,
sortable=True
)
gb.configure_column(
"text",
width=500,
wrapText=True,
autoHeight=True,
editable=False
)
gb.configure_column(
"ground_truth",
editable=True,
width=300,
cellStyle={'backgroundColor': '#fffde7'}
)
gb.configure_selection(
selection_mode='multiple',
use_checkbox=True,
pre_selected_rows=pre_selected_rows
)
gb.configure_pagination(paginationAutoPageSize=False, paginationPageSize=10)
grid_response = AgGrid(
df,
gridOptions=gb.build(),
update_mode=GridUpdateMode.MODEL_CHANGED | GridUpdateMode.SELECTION_CHANGED,
data_return_mode=DataReturnMode.FILTERED_AND_SORTED,
fit_columns_on_grid_load=False,
theme='streamlit',
height=400,
key='annotation_grid'
)
# Update session state with edited data
st.session_state.dataset = pd.DataFrame(grid_response['data'])
# Update selected indices from grid selection
selected_rows = grid_response.get('selected_rows', [])
if selected_rows is not None and len(selected_rows) > 0:
# Get indices of selected rows
selected_df = pd.DataFrame(selected_rows)
if not selected_df.empty:
# Match selected rows back to original indices
st.session_state.selected_indices = selected_df.index.tolist()
else:
st.session_state.selected_indices = []
st.divider()
# Save/Export section
st.subheader("๐Ÿ’พ Save Annotated Data")
col1, col2, col3 = st.columns([2, 1, 1])
with col1:
save_filename = st.text_input(
"Filename",
value="annotated_data.csv",
help="Enter filename (.csv or .json)"
)
with col2:
if st.button("๐Ÿ’พ Save to File", use_container_width=True):
if save_annotated_data(
st.session_state.dataset,
st.session_state.selected_indices,
save_filename
):
st.success(f"โœ… Saved to {save_filename}")
with col3:
# Download button
save_df = st.session_state.dataset.copy()
save_df['_selected'] = save_df.index.isin(st.session_state.selected_indices)
csv_data = save_df.to_csv(index=False)
st.download_button(
"๐Ÿ“ฅ Download CSV",
csv_data,
file_name="annotated_data.csv",
mime="text/csv",
use_container_width=True
)
st.divider()
# Data statistics
st.subheader("๐Ÿ“Š Data Statistics")
total = len(st.session_state.dataset)
annotated = (st.session_state.dataset['ground_truth'].astype(str) != '').sum()
selected_count = len(st.session_state.selected_indices)
# Calculate train/val split preview
selected_df = st.session_state.dataset.iloc[st.session_state.selected_indices] if st.session_state.selected_indices else pd.DataFrame()
selected_annotated = selected_df[selected_df['ground_truth'].astype(str) != ''] if not selected_df.empty else pd.DataFrame()
if len(selected_annotated) >= 2:
train_df, val_df = stratified_train_val_split(
selected_annotated,
train_ratio=st.session_state.train_test_split
)
train_size = len(train_df)
val_size = len(val_df)
else:
train_size = 0
val_size = 0
col1, col2, col3, col4 = st.columns(4)
with col1:
st.metric("Total Samples", total)
with col2:
st.metric("Annotated", f"{annotated}/{total}")
with col3:
st.metric("Selected", selected_count, help="Rows selected for training/validation")
with col4:
st.metric("Train/Val", f"{train_size}/{val_size}", help="Stratified split of selected & annotated rows")
# Selection guidance
if selected_count == 0:
st.info("๐Ÿ’ก Select rows using checkboxes to include them in training/validation.")
elif len(selected_annotated) < 2:
st.warning("โš ๏ธ Please select at least 2 annotated rows for training.")
# Stratification preview
if len(selected_annotated) >= 2:
with st.expander("๐Ÿ“ˆ Stratification Preview"):
# Show distribution of ground_truth patterns
pattern_counts = selected_annotated['ground_truth'].apply(
lambda x: str(x)[:30] + '...' if len(str(x)) > 30 else str(x)
).value_counts()
st.markdown("**Ground Truth Pattern Distribution:**")
st.bar_chart(pattern_counts)
st.caption(f"Training: {train_size} samples, Validation: {val_size} samples")
# Sample data preview
with st.expander("๐Ÿ“‹ Sample Preview"):
st.dataframe(
st.session_state.dataset.head(5),
use_container_width=True
)
def render_optimization_tab():
"""Render the optimization tab."""
st.header("๐Ÿงฌ GEPA Optimization")
if st.session_state.dataset is None:
st.warning("โš ๏ธ Please upload and annotate data first.")
return
df = st.session_state.dataset
selected_indices = st.session_state.get('selected_indices', [])
# Use selected rows if available, otherwise fall back to all annotated rows
if selected_indices:
selected_df = df.iloc[selected_indices]
annotated_df = selected_df[selected_df['ground_truth'].astype(str) != '']
use_selection = True
else:
annotated_df = df[df['ground_truth'].astype(str) != '']
use_selection = False
if len(annotated_df) < 2:
if use_selection:
st.warning("โš ๏ธ Please select and annotate at least 2 samples in the Data Ingestion tab.")
else:
st.warning("โš ๏ธ Please annotate at least 2 samples or select rows for training.")
return
# Stratified split
train_df, val_df = stratified_train_val_split(
annotated_df,
train_ratio=st.session_state.train_test_split
)
col1, col2, col3 = st.columns(3)
with col1:
st.info(f"๐Ÿ“š Training samples: {len(train_df)}")
with col2:
st.info(f"๐Ÿงช Validation samples: {len(val_df)}")
with col3:
if use_selection:
st.success("โœ… Using selected rows")
else:
st.warning("โš ๏ธ Using all annotated rows")
# Optimization controls
col1, col2, col3 = st.columns([1, 1, 2])
with col1:
run_button = st.button(
"๐Ÿš€ Run Optimization",
type="primary",
use_container_width=True
)
with col2:
if st.button("๐Ÿ”„ Reset Results", use_container_width=True):
st.session_state.optimized_program = None
st.session_state.optimization_history = []
st.rerun()
if run_button:
if not setup_dspy():
return
# Prepare training set
trainset = [
dspy.Example(
raw_text=row['text'],
ground_truth=row['ground_truth']
).with_inputs('raw_text')
for _, row in train_df.iterrows()
]
valset = [
dspy.Example(
raw_text=row['text'],
ground_truth=row['ground_truth']
).with_inputs('raw_text')
for _, row in val_df.iterrows()
]
# Progress tracking
progress_bar = st.progress(0)
status_text = st. empty()
try:
with st.spinner("๐Ÿงฌ GEPA is evolving regex patterns..."):
status_text.text("Initializing optimizer...")
optimizer = GEPA(
metric=create_regex_metric(st.session_state.regex_flags),
num_iterations=st. session_state.gepa_config['num_iterations'],
num_candidates=st.session_state.gepa_config['num_candidates'],
)
progress_bar.progress(20)
status_text.text("Creating initial program...")
program = RegexGenerator(
doc=st.session_state.prompts['system_instruction'],
output_desc=st. session_state.prompts['output_description']
)
progress_bar.progress(40)
status_text.text("Running optimization...")
optimized = optimizer.compile(
program,
trainset=trainset,
)
progress_bar.progress(80)
status_text.text("Evaluating on validation set...")
# Evaluate on validation set
metric_fn = create_regex_metric(st.session_state.regex_flags)
val_scores = []
for example in valset:
pred = optimized(raw_text=example. raw_text)
result = metric_fn(example, pred)
val_scores.append(result. score)
avg_score = sum(val_scores) / len(val_scores) if val_scores else 0
progress_bar. progress(100)
status_text.text("Complete!")
st.session_state. optimized_program = optimized
st.session_state.optimization_history.append({
'score': avg_score,
'prompt': optimized.predictor.signature.__doc__,
'timestamp': pd.Timestamp.now()
})
st. success(f"โœ… Optimization Complete! Validation Score: {avg_score:.2%}")
except Exception as e:
st.error(f"Optimization failed: {e}")
return
# Display results
if st. session_state.optimized_program:
st.subheader("๐Ÿ“Š Results")
with st.expander("๐Ÿ” Evolved Prompt", expanded=True):
st.code(
st.session_state.optimized_program.predictor. signature.__doc__,
language="text"
)
# Optimization history
if st.session_state.optimization_history:
with st.expander("๐Ÿ“ˆ Optimization History"):
history_df = pd. DataFrame(st.session_state. optimization_history)
st.dataframe(history_df, use_container_width=True)
def render_testing_tab():
"""Render the testing tab."""
st.header("๐Ÿ” Test & Validate")
if st.session_state.optimized_program is None:
st. warning("โš ๏ธ Please run optimization first.")
return
# Single test
st.subheader("๐Ÿงช Single Test")
test_input = st.text_area(
"Enter test text",
height=100,
placeholder="Paste text here to extract regex pattern..."
)
col1, col2 = st.columns([1, 3])
with col1:
test_button = st.button("โ–ถ๏ธ Generate & Run", type="primary")
if test_button and test_input:
if not setup_dspy():
return
with st.spinner("Generating regex... "):
try:
result = st.session_state.optimized_program(raw_text=test_input)
pattern = result.regex_pattern
st.code(f"Generated Regex: {pattern}", language="regex")
# Compile and test
flags = 0
for flag in st.session_state.regex_flags:
flags |= getattr(re, flag, 0)
compiled = re.compile(pattern, flags)
matches = compiled.findall(test_input)
if matches:
st.success(f"โœ… Found {len(matches)} match(es):")
for i, match in enumerate(matches, 1):
st.markdown(f"**Match {i}:** `{match}`")
# Highlight matches in text
highlighted = test_input
for match in matches:
highlighted = highlighted.replace(
match,
f"**: green[{match}]**"
)
st.markdown("**Highlighted text:**")
st.markdown(highlighted)
else:
st. warning("No matches found.")
except re.error as e:
st.error(f"Invalid regex generated: {e}")
except Exception as e:
st.error(f"Error: {e}")
st.divider()
# Batch testing
st. subheader("๐Ÿ“‹ Batch Testing")
batch_file = st.file_uploader(
"Upload test data (CSV with 'text' column)",
type=["csv"],
key="batch_test"
)
if batch_file:
test_df = pd. read_csv(batch_file)
if 'text' not in test_df. columns:
st.error("CSV must have 'text' column.")
return
if st.button("๐Ÿš€ Run Batch Test"):
if not setup_dspy():
return
results = []
progress = st.progress(0)
for i, row in test_df.iterrows():
try:
result = st.session_state.optimized_program(raw_text=row['text'])
pattern = result.regex_pattern
flags = 0
for flag in st.session_state. regex_flags:
flags |= getattr(re, flag, 0)
match = re.search(pattern, row['text'], flags)
extracted = match.group(0) if match else ""
results.append({
'text': row['text'][: 100] + '...' if len(row['text']) > 100 else row['text'],
'pattern': pattern,
'extracted': extracted,
'success': bool(match)
})
except Exception as e:
results.append({
'text': row['text'][:100] + '...',
'pattern': 'ERROR',
'extracted': str(e),
'success': False
})
progress. progress((i + 1) / len(test_df))
results_df = pd. DataFrame(results)
# Summary metrics
success_rate = results_df['success']. mean()
col1, col2 = st.columns(2)
with col1:
st.metric("Success Rate", f"{success_rate:.1%}")
with col2:
st.metric("Total Tests", len(results_df))
# Results table
st.dataframe(results_df, use_container_width=True)
# Download results
csv = results_df. to_csv(index=False)
st.download_button(
"๐Ÿ“ฅ Download Results",
csv,
"batch_test_results. csv",
"text/csv"
)
# --- Main Application ---
def main():
render_sidebar()
st.title("๐Ÿงฌ GEPA Regex Optimizer")
st.caption("Automated regex generation with DSPy and evolutionary optimization")
tab1, tab2, tab3 = st.tabs([
"๐Ÿ“ฅ Data Ingestion",
"๐Ÿงฌ Optimization",
"๐Ÿ” Testing"
])
with tab1:
render_data_ingestion_tab()
with tab2:
render_optimization_tab()
with tab3:
render_testing_tab()
# Footer
st.divider()
st.caption(
"Built with Streamlit, DSPy, and GEPA | "
"Configuration is auto-saved in the sidebar"
)
if __name__ == "__main__":
main()