Antigravity Agent commited on
Commit ·
3c969ac
1
Parent(s): be2c8ad
feat(ui): complete Senior UI overhaul with glassmorphism and signal intelligence
Browse files
app.py
CHANGED
|
@@ -6,6 +6,7 @@ import numpy as np
|
|
| 6 |
from PIL import Image
|
| 7 |
import tempfile
|
| 8 |
import json
|
|
|
|
| 9 |
|
| 10 |
# Import consolidated modules
|
| 11 |
from ocr_module import MVM2OCREngine
|
|
@@ -19,105 +20,188 @@ from image_enhancing import ImageEnhancer
|
|
| 19 |
ocr_engine = MVM2OCREngine()
|
| 20 |
enhancer = ImageEnhancer(sigma=1.2)
|
| 21 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 22 |
def process_mvm2_pipeline(image, auto_enhance):
|
| 23 |
if image is None:
|
| 24 |
-
return
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 25 |
|
| 26 |
-
# 1. Preprocessing
|
| 27 |
-
if auto_enhance:
|
| 28 |
-
enhanced_img_np, meta = enhancer.enhance(image)
|
| 29 |
-
# Save temp image for OCR
|
| 30 |
-
temp_img_path = os.path.join(tempfile.gettempdir(), 'enhanced_input.png')
|
| 31 |
-
cv2.imwrite(temp_img_path, enhanced_img_np)
|
| 32 |
-
else:
|
| 33 |
-
# Save original PIL image
|
| 34 |
-
temp_img_path = os.path.join(tempfile.gettempdir(), 'original_input.png')
|
| 35 |
-
image.save(temp_img_path)
|
| 36 |
-
meta = {'metrics': {'initial_contrast': 0}}
|
| 37 |
-
|
| 38 |
# 2. OCR Extraction
|
| 39 |
ocr_results = ocr_engine.process_image(temp_img_path)
|
| 40 |
latex_text = ocr_results['latex_output']
|
| 41 |
ocr_conf = ocr_results['weighted_confidence']
|
| 42 |
|
| 43 |
-
if 'No math detected' in latex_text:
|
| 44 |
-
return f'OCR Failure: {latex_text}', None, None
|
| 45 |
-
|
| 46 |
# 3. Multi-Agent Reasoning
|
| 47 |
agent_responses = run_agent_orchestrator(latex_text)
|
| 48 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 49 |
# 4. Consensus Fusion
|
| 50 |
consensus_result = evaluate_consensus(agent_responses, ocr_confidence=ocr_conf)
|
| 51 |
|
| 52 |
-
#
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 53 |
reports = generate_mvm2_report(consensus_result, latex_text, ocr_conf)
|
| 54 |
-
md_report =
|
| 55 |
-
json_report = json.loads(reports['json'])
|
| 56 |
|
| 57 |
-
# 6. Export to PDF
|
| 58 |
pdf_path = os.path.join(tempfile.gettempdir(), f'MVM2_Report_{reports["report_id"]}.pdf')
|
| 59 |
-
export_to_pdf(
|
| 60 |
|
| 61 |
-
return
|
| 62 |
|
| 63 |
-
#
|
| 64 |
-
|
| 65 |
-
.
|
| 66 |
-
|
| 67 |
-
|
| 68 |
-
|
| 69 |
-
|
| 70 |
-
|
| 71 |
-
|
| 72 |
-
|
| 73 |
-
|
| 74 |
-
margin-bottom: 20px;
|
| 75 |
-
}
|
| 76 |
-
.report-area {
|
| 77 |
-
background-color: #f9f9f9;
|
| 78 |
-
padding: 15px;
|
| 79 |
-
border-radius: 8px;
|
| 80 |
-
border: 1px solid #ddd;
|
| 81 |
-
}
|
| 82 |
-
"""
|
| 83 |
-
|
| 84 |
-
with gr.Blocks(css=custom_css, title='MVM2: Math Verification & Multi-Signal Consensus') as demo:
|
| 85 |
-
gr.Markdown(
|
| 86 |
-
"""
|
| 87 |
-
<div class="mvm2-header">
|
| 88 |
-
<h1>MVM2: Neuro-Symbolic Math Verification</h1>
|
| 89 |
-
<p>Adaptive Multi-Signal Consensus for Handwritten Mathematical Equation Verification</p>
|
| 90 |
-
</div>
|
| 91 |
-
"""
|
| 92 |
-
)
|
| 93 |
|
| 94 |
with gr.Row():
|
| 95 |
-
|
| 96 |
-
|
| 97 |
-
|
| 98 |
-
|
|
|
|
|
|
|
| 99 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 100 |
with gr.Column(scale=2):
|
| 101 |
-
|
| 102 |
-
|
| 103 |
-
|
| 104 |
-
|
| 105 |
-
|
| 106 |
-
|
|
|
|
| 107 |
|
| 108 |
-
|
| 109 |
-
""
|
| 110 |
-
|
| 111 |
-
|
| 112 |
-
|
| 113 |
-
|
| 114 |
-
|
| 115 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 116 |
|
| 117 |
run_btn.click(
|
| 118 |
fn=process_mvm2_pipeline,
|
| 119 |
inputs=[input_img, enhance_toggle],
|
| 120 |
-
outputs=[
|
| 121 |
)
|
| 122 |
|
| 123 |
if __name__ == "__main__":
|
|
|
|
| 6 |
from PIL import Image
|
| 7 |
import tempfile
|
| 8 |
import json
|
| 9 |
+
import re
|
| 10 |
|
| 11 |
# Import consolidated modules
|
| 12 |
from ocr_module import MVM2OCREngine
|
|
|
|
| 20 |
ocr_engine = MVM2OCREngine()
|
| 21 |
enhancer = ImageEnhancer(sigma=1.2)
|
| 22 |
|
| 23 |
+
# Load custom CSS
|
| 24 |
+
with open("theme.css", "r") as f:
|
| 25 |
+
css_content = f.read()
|
| 26 |
+
|
| 27 |
+
def create_gauge(label, value, color="#6366f1"):
|
| 28 |
+
"""Generates an animated SVG circular gauge."""
|
| 29 |
+
percentage = max(0, min(100, value * 100))
|
| 30 |
+
dash_offset = 251.2 * (1 - percentage / 100)
|
| 31 |
+
return f"""
|
| 32 |
+
<div class="gauge-container">
|
| 33 |
+
<svg width="100" height="100" viewBox="0 0 100 100">
|
| 34 |
+
<circle class="circle-bg" cx="50" cy="50" r="40" />
|
| 35 |
+
<circle class="circle-progress" cx="50" cy="50" r="40"
|
| 36 |
+
stroke="{color}" stroke-dasharray="251.2"
|
| 37 |
+
stroke-dashoffset="{dash_offset}"
|
| 38 |
+
style="filter: drop-shadow(0 0 5px {color}88);"/>
|
| 39 |
+
<text x="50" y="55" text-anchor="middle" font-size="18" font-weight="bold" fill="white">{int(percentage)}%</text>
|
| 40 |
+
</svg>
|
| 41 |
+
<div style="font-size: 0.8em; color: #94a3b8; font-weight: 500;">{label}</div>
|
| 42 |
+
</div>
|
| 43 |
+
"""
|
| 44 |
+
|
| 45 |
+
def format_step_viewer(consensus_result):
|
| 46 |
+
"""Formats the Reasoning Trace with Step-Level Consensus highlights."""
|
| 47 |
+
html = '<div style="display: flex; flex-direction: column; gap: 12px;">'
|
| 48 |
+
|
| 49 |
+
# We aggregate steps from all agents for a collective view
|
| 50 |
+
agent_data = consensus_result.get("detail_scores", [])
|
| 51 |
+
|
| 52 |
+
for agent in agent_data:
|
| 53 |
+
# Simulate step-level analysis for UI purposes:
|
| 54 |
+
# In a real system, we'd have Score_j per step. Here we use the agent's overall score.
|
| 55 |
+
score = agent["Score_j"]
|
| 56 |
+
status_class = "step-valid" if score >= 0.7 else "step-warning"
|
| 57 |
+
icon = "✅" if score >= 0.7 else "⚠️"
|
| 58 |
+
glow_style = "box-shadow: 0 0 10px rgba(16, 185, 129, 0.2);" if score >= 0.7 else "box-shadow: 0 0 10px rgba(245, 158, 11, 0.2);"
|
| 59 |
+
|
| 60 |
+
# Get matching agent response for trace
|
| 61 |
+
# (This assumes agent_responses were passed in or stored)
|
| 62 |
+
# For the UI, we'll just show the representative trace from valid agents
|
| 63 |
+
if not agent["is_hallucinating"] or score > 0.4:
|
| 64 |
+
html += f"""
|
| 65 |
+
<div class="glass-card reasoning-step {status_class}" style="{glow_style}">
|
| 66 |
+
<div style="display: flex; justify-content: space-between; align-items: center; margin-bottom: 8px;">
|
| 67 |
+
<span class="monospace" style="color: #6366f1; font-weight: 600;">{agent['agent']} Reasoning Path</span>
|
| 68 |
+
<span style="font-size: 0.8em; background: rgba(0,0,0,0.3); padding: 2px 8px; border-radius: 4px;">Consensus: {score:.2f} {icon}</span>
|
| 69 |
+
</div>
|
| 70 |
+
<div style="font-size: 0.9em; line-height: 1.6; color: #cbd5e1;">
|
| 71 |
+
{"<br>".join([f"• {step}" for step in agent.get('reasoning_trace', ['Processing...'])])}
|
| 72 |
+
</div>
|
| 73 |
+
</div>
|
| 74 |
+
"""
|
| 75 |
+
html += "</div>"
|
| 76 |
+
return html
|
| 77 |
+
|
| 78 |
def process_mvm2_pipeline(image, auto_enhance):
|
| 79 |
if image is None:
|
| 80 |
+
return None, "Please upload an image.", None, "", None, ""
|
| 81 |
+
|
| 82 |
+
# 1. Preprocessing & Preview
|
| 83 |
+
enhanced_img_np, meta = enhancer.enhance(image)
|
| 84 |
+
temp_img_path = os.path.join(tempfile.gettempdir(), 'input_processed.png')
|
| 85 |
+
cv2.imwrite(temp_img_path, enhanced_img_np)
|
| 86 |
+
|
| 87 |
+
preview_img = Image.fromarray(cv2.cvtColor(enhanced_img_np, cv2.COLOR_BGR2RGB))
|
| 88 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 89 |
# 2. OCR Extraction
|
| 90 |
ocr_results = ocr_engine.process_image(temp_img_path)
|
| 91 |
latex_text = ocr_results['latex_output']
|
| 92 |
ocr_conf = ocr_results['weighted_confidence']
|
| 93 |
|
|
|
|
|
|
|
|
|
|
| 94 |
# 3. Multi-Agent Reasoning
|
| 95 |
agent_responses = run_agent_orchestrator(latex_text)
|
| 96 |
|
| 97 |
+
# Attach traces back to detail_scores for UI formatting
|
| 98 |
+
for i, res in enumerate(agent_responses):
|
| 99 |
+
agent_responses[i]["response"]["agent_id"] = i # tag
|
| 100 |
+
|
| 101 |
# 4. Consensus Fusion
|
| 102 |
consensus_result = evaluate_consensus(agent_responses, ocr_confidence=ocr_conf)
|
| 103 |
|
| 104 |
+
# Map traces to detail_scores for UI
|
| 105 |
+
for i, score_data in enumerate(consensus_result["detail_scores"]):
|
| 106 |
+
# Match by agent name
|
| 107 |
+
for res in agent_responses:
|
| 108 |
+
if res["agent"] == score_data["agent"]:
|
| 109 |
+
consensus_result["detail_scores"][i]["reasoning_trace"] = res["response"].get("Reasoning Trace", [])
|
| 110 |
+
break
|
| 111 |
+
|
| 112 |
+
# 5. Gauges & UI Elements
|
| 113 |
+
avg_v_sym = np.mean([s["V_sym"] for s in consensus_result["detail_scores"]])
|
| 114 |
+
avg_l_logic = np.mean([s["L_logic"] for s in consensus_result["detail_scores"]])
|
| 115 |
+
avg_c_clf = np.mean([s["C_clf"] for s in consensus_result["detail_scores"]])
|
| 116 |
+
|
| 117 |
+
gauges_html = f"""
|
| 118 |
+
<div class="signal-panel">
|
| 119 |
+
{create_gauge("Symbolic", avg_v_sym, "#10b981")}
|
| 120 |
+
{create_gauge("Logic", avg_l_logic, "#6366f1")}
|
| 121 |
+
{create_gauge("Classifier", avg_c_clf, "#8b5cf6")}
|
| 122 |
+
</div>
|
| 123 |
+
"""
|
| 124 |
+
|
| 125 |
+
# Final Calibration Bar
|
| 126 |
+
winner = consensus_result["winning_score"]
|
| 127 |
+
calibrated_conf = winner * (0.9 + 0.1 * ocr_conf)
|
| 128 |
+
conf_bar = f"""
|
| 129 |
+
<div style="margin-top: 20px;">
|
| 130 |
+
<div style="display: flex; justify-content: space-between; margin-bottom: 8px;">
|
| 131 |
+
<span style="font-weight: 600; color: #94a3b8;">Final Confidence Calibration</span>
|
| 132 |
+
<span style="color: #10b981; font-weight: bold;">{calibrated_conf:.3f}</span>
|
| 133 |
+
</div>
|
| 134 |
+
<div style="width: 100%; bg: rgba(255,255,255,0.05); height: 8px; border-radius: 4px; overflow: hidden;">
|
| 135 |
+
<div style="width: {min(100, calibrated_conf*50)}%; background: linear-gradient(90deg, #6366f1 0%, #10b981 100%); height: 100%; transition: width 1s ease;"></div>
|
| 136 |
+
</div>
|
| 137 |
+
</div>
|
| 138 |
+
"""
|
| 139 |
+
|
| 140 |
+
# 6. Report & PDF
|
| 141 |
reports = generate_mvm2_report(consensus_result, latex_text, ocr_conf)
|
| 142 |
+
md_report = format_step_viewer(consensus_result)
|
|
|
|
| 143 |
|
|
|
|
| 144 |
pdf_path = os.path.join(tempfile.gettempdir(), f'MVM2_Report_{reports["report_id"]}.pdf')
|
| 145 |
+
export_to_pdf(json.loads(reports['json']), pdf_path)
|
| 146 |
|
| 147 |
+
return preview_img, latex_text, gauges_html, conf_bar, md_report, pdf_path
|
| 148 |
|
| 149 |
+
# Build Interface
|
| 150 |
+
with gr.Blocks(css=css_content, title="MVM²: Senior UI AI Dashboard") as demo:
|
| 151 |
+
with gr.Row(elem_id="header-row"):
|
| 152 |
+
gr.Markdown(
|
| 153 |
+
"""
|
| 154 |
+
<div style="text-align: center; padding: 20px 0;">
|
| 155 |
+
<h1 style="font-size: 2.5em; margin-bottom: 0;">MVM² <span style="color: #6366f1;">Neuro-Symbolic</span></h1>
|
| 156 |
+
<p style="color: #94a3b8; font-size: 1.1em; margin-top: 8px;">High-Fidelity Mathematical Verification & Consensus Dashboard</p>
|
| 157 |
+
</div>
|
| 158 |
+
"""
|
| 159 |
+
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 160 |
|
| 161 |
with gr.Row():
|
| 162 |
+
# --- LEFT PANEL: Upload & Preview ---
|
| 163 |
+
with gr.Column(scale=1, variant="panel"):
|
| 164 |
+
gr.Markdown("### 📤 Input Intelligence")
|
| 165 |
+
input_img = gr.Image(type="pil", label="Capture Solution", elem_classes="glass-card")
|
| 166 |
+
enhance_toggle = gr.Checkbox(label="Enable Opti-Scan Preprocessing", value=True)
|
| 167 |
+
run_btn = gr.Button("INITIALIZE VERIFICATION", variant="primary", elem_classes="download-btn")
|
| 168 |
|
| 169 |
+
gr.Markdown("#### 🔍 Preprocessing Preview")
|
| 170 |
+
preview_output = gr.Image(label="Enhanced Signal", interactive=False, elem_classes="preview-img")
|
| 171 |
+
|
| 172 |
+
# --- CENTER STAGE: Canvas ---
|
| 173 |
with gr.Column(scale=2):
|
| 174 |
+
gr.Markdown("### 🎨 MVM² Verification Canvas")
|
| 175 |
+
with gr.Box(elem_classes="glass-card"):
|
| 176 |
+
canvas_latex = gr.Textbox(label="Canonical LaTeX Transcription", lines=2, interactive=False, elem_classes="monospace")
|
| 177 |
+
calib_bar_html = gr.HTML()
|
| 178 |
+
|
| 179 |
+
gr.Markdown("### 🪜 Dynamic Reasoning Trace")
|
| 180 |
+
trace_html = gr.HTML()
|
| 181 |
|
| 182 |
+
# --- RIGHT PANEL: Signal Intel ---
|
| 183 |
+
with gr.Column(scale=1, variant="panel"):
|
| 184 |
+
gr.Markdown("### ⚡ Signal Intelligence")
|
| 185 |
+
with gr.Box(elem_classes="glass-card"):
|
| 186 |
+
signal_gauges = gr.HTML()
|
| 187 |
+
|
| 188 |
+
gr.Markdown("### 📄 Educational Assessment")
|
| 189 |
+
download_btn = gr.File(label="Download Diagnostic PDF", elem_classes="download-btn")
|
| 190 |
+
|
| 191 |
+
with gr.Box(elem_classes="glass-card", style="margin-top: 20px; border-left: 4px solid #6366f1;"):
|
| 192 |
+
gr.Markdown(
|
| 193 |
+
"""
|
| 194 |
+
**System Status**
|
| 195 |
+
- Pix2Text VLM: `Online`
|
| 196 |
+
- SymPy Core: `1.12.0`
|
| 197 |
+
- Consensus: `4-Agent parallel`
|
| 198 |
+
"""
|
| 199 |
+
)
|
| 200 |
|
| 201 |
run_btn.click(
|
| 202 |
fn=process_mvm2_pipeline,
|
| 203 |
inputs=[input_img, enhance_toggle],
|
| 204 |
+
outputs=[preview_output, canvas_latex, signal_gauges, calib_bar_html, trace_html, download_btn]
|
| 205 |
)
|
| 206 |
|
| 207 |
if __name__ == "__main__":
|
theme.css
ADDED
|
@@ -0,0 +1,130 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
/* MVM² Senior UI Design System - Linear/Vercel Inspired */
|
| 2 |
+
@import url('https://fonts.googleapis.com/css2?family=Inter:wght@300;400;500;600;700&family=JetBrains+Mono:wght@400;500&display=swap');
|
| 3 |
+
|
| 4 |
+
:root {
|
| 5 |
+
--bg-color: #0f172a;
|
| 6 |
+
--card-bg: rgba(30, 41, 59, 0.7);
|
| 7 |
+
--primary: #6366f1;
|
| 8 |
+
--primary-glow: rgba(99, 102, 241, 0.3);
|
| 9 |
+
--success: #10b981;
|
| 10 |
+
--success-glow: rgba(16, 185, 129, 0.2);
|
| 11 |
+
--warning: #f59e0b;
|
| 12 |
+
--warning-glow: rgba(245, 158, 11, 0.2);
|
| 13 |
+
--border: rgba(255, 255, 255, 0.1);
|
| 14 |
+
--text-main: #f8f9fa;
|
| 15 |
+
--text-muted: #94a3b8;
|
| 16 |
+
}
|
| 17 |
+
|
| 18 |
+
body, .gradio-container {
|
| 19 |
+
background-color: var(--bg-color) !important;
|
| 20 |
+
color: var(--text-main) !important;
|
| 21 |
+
font-family: 'Inter', sans-serif !important;
|
| 22 |
+
}
|
| 23 |
+
|
| 24 |
+
/* Glassmorphism Cards */
|
| 25 |
+
.glass-card {
|
| 26 |
+
background: var(--card-bg);
|
| 27 |
+
backdrop-filter: blur(12px);
|
| 28 |
+
-webkit-backdrop-filter: blur(12px);
|
| 29 |
+
border: 1px solid var(--border);
|
| 30 |
+
border-radius: 16px;
|
| 31 |
+
padding: 24px;
|
| 32 |
+
transition: all 0.3s ease;
|
| 33 |
+
}
|
| 34 |
+
|
| 35 |
+
.glass-card:hover {
|
| 36 |
+
border-color: var(--primary);
|
| 37 |
+
box-shadow: 0 0 20px var(--primary-glow);
|
| 38 |
+
}
|
| 39 |
+
|
| 40 |
+
/* Typography Enhancements */
|
| 41 |
+
h1 { font-weight: 700; letter-spacing: -0.025em; color: #fff; }
|
| 42 |
+
h2 { font-weight: 600; color: var(--text-main); }
|
| 43 |
+
.monospace { font-family: 'JetBrains Mono', monospace !important; font-size: 0.9em; }
|
| 44 |
+
|
| 45 |
+
/* Custom Accordion Styling */
|
| 46 |
+
.reasoning-step {
|
| 47 |
+
margin-bottom: 8px;
|
| 48 |
+
padding: 12px;
|
| 49 |
+
border-radius: 8px;
|
| 50 |
+
background: rgba(255, 255, 255, 0.03);
|
| 51 |
+
border-left: 4px solid var(--text-muted);
|
| 52 |
+
transition: transform 0.2s ease;
|
| 53 |
+
}
|
| 54 |
+
|
| 55 |
+
.step-valid {
|
| 56 |
+
border-left-color: var(--success);
|
| 57 |
+
background: var(--success-glow);
|
| 58 |
+
box-shadow: inset 0 0 10px rgba(16, 185, 129, 0.1);
|
| 59 |
+
}
|
| 60 |
+
|
| 61 |
+
.step-warning {
|
| 62 |
+
border-left-color: var(--warning);
|
| 63 |
+
background: var(--warning-glow);
|
| 64 |
+
}
|
| 65 |
+
|
| 66 |
+
/* Animated Gauges */
|
| 67 |
+
.gauge-container {
|
| 68 |
+
display: flex;
|
| 69 |
+
flex-direction: column;
|
| 70 |
+
align-items: center;
|
| 71 |
+
gap: 12px;
|
| 72 |
+
}
|
| 73 |
+
|
| 74 |
+
.circle-bg { fill: none; stroke: rgba(255,255,255,0.05); stroke-width: 8; }
|
| 75 |
+
.circle-progress {
|
| 76 |
+
fill: none;
|
| 77 |
+
stroke-width: 8;
|
| 78 |
+
stroke-linecap: round;
|
| 79 |
+
transition: stroke-dashoffset 1s ease-in-out;
|
| 80 |
+
}
|
| 81 |
+
|
| 82 |
+
@keyframes dash {
|
| 83 |
+
from { stroke-dashoffset: 251.2; }
|
| 84 |
+
}
|
| 85 |
+
|
| 86 |
+
/* Signal Intel Panel */
|
| 87 |
+
.signal-panel {
|
| 88 |
+
display: grid;
|
| 89 |
+
grid-template-columns: repeat(3, 1fr);
|
| 90 |
+
gap: 16px;
|
| 91 |
+
margin-top: 20px;
|
| 92 |
+
}
|
| 93 |
+
|
| 94 |
+
/* Download Button Animation */
|
| 95 |
+
.download-btn {
|
| 96 |
+
background: linear-gradient(135deg, var(--primary) 0%, #4338ca 100%) !important;
|
| 97 |
+
border: none !important;
|
| 98 |
+
color: white !important;
|
| 99 |
+
font-weight: 600 !important;
|
| 100 |
+
transition: all 0.3s cubic-bezier(0.4, 0, 0.2, 1) !important;
|
| 101 |
+
}
|
| 102 |
+
|
| 103 |
+
.download-btn:hover {
|
| 104 |
+
transform: translateY(-2px);
|
| 105 |
+
box-shadow: 0 10px 25px -5px var(--primary-glow);
|
| 106 |
+
filter: brightness(1.1);
|
| 107 |
+
}
|
| 108 |
+
|
| 109 |
+
/* Preprocessing Preview */
|
| 110 |
+
.preview-img img {
|
| 111 |
+
border-radius: 12px;
|
| 112 |
+
border: 1px solid var(--border);
|
| 113 |
+
filter: drop-shadow(0 4px 6px rgba(0,0,0,0.3));
|
| 114 |
+
}
|
| 115 |
+
|
| 116 |
+
/* Gradio Overrides */
|
| 117 |
+
.stButton button { border-radius: 10px !important; }
|
| 118 |
+
.stMarkdown { color: var(--text-main) !important; }
|
| 119 |
+
input, textarea {
|
| 120 |
+
background: rgba(15, 23, 42, 0.6) !important;
|
| 121 |
+
border: 1px solid var(--border) !important;
|
| 122 |
+
color: white !important;
|
| 123 |
+
border-radius: 8px !important;
|
| 124 |
+
}
|
| 125 |
+
input:focus, textarea:focus { border-color: var(--primary) !important; }
|
| 126 |
+
|
| 127 |
+
/* Responsive Adjustments */
|
| 128 |
+
@media (max-width: 768px) {
|
| 129 |
+
.signal-panel { grid-template-columns: 1fr; }
|
| 130 |
+
}
|