Spaces:
No application file
No application file
| # HF Spaces / Gradio app: Vochi CRM call logs + AI analysis | |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ | |
| # How to deploy (short): | |
| # 1) Create a new Space (Python + Gradio). | |
| # 2) Add a file named `app.py` with THIS code. | |
| # 3) Add a file named `requirements.txt` with the lines below. | |
| # 4) In the Space β Settings β Repository secrets, add: | |
| # - VOCHI_BASE_URL (e.g. https://crm.vochi.by/api) | |
| # - VOCHI_CLIENT_ID (client id string) | |
| # - GOOGLE_API_KEY (API key) | |
| # | |
| # UI language: English. | |
| from __future__ import annotations | |
| import os | |
| import json | |
| import datetime as _dt | |
| from typing import List, Tuple, Optional | |
| import requests | |
| import pandas as pd | |
| import numpy as np | |
| import gradio as gr | |
| try: | |
| # New Google Gemini client library | |
| from google import genai # type: ignore | |
| _HAS_GENAI = True | |
| except Exception: | |
| genai = None | |
| _HAS_GENAI = False | |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ | |
| # Config | |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ | |
| BASE_URL = os.environ.get("VOCHI_BASE_URL", "https://crm.vochi.by/api") | |
| CLIENT_ID = os.environ.get("VOCHI_CLIENT_ID") | |
| # If your API needs auth, fill it here (or via VOCHI_BEARER in Secrets) | |
| _AUTH_TOKEN = os.environ.get("VOCHI_BEARER", "").strip() | |
| AUTH_HEADERS = { | |
| "Accept": "audio/*,application/json;q=0.9,*/*;q=0.8", | |
| **({"Authorization": f"Bearer {_AUTH_TOKEN}"} if _AUTH_TOKEN else {}), | |
| } | |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ | |
| # Vochi API helpers | |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ | |
| def fetch_calllogs(date_str: str): | |
| """Get list of calls for a given date (YYYY-MM-DD).""" | |
| r = requests.get( | |
| f"{BASE_URL}/calllogs", | |
| params={"start": date_str, "end": date_str, "clientId": CLIENT_ID}, | |
| headers=AUTH_HEADERS, | |
| timeout=60, | |
| ) | |
| r.raise_for_status() | |
| data = r.json() | |
| if isinstance(data, dict): | |
| return data.get("data", data) | |
| return data | |
| def fetch_mp3_by_unique_id(unique_id: str) -> Tuple[str, str]: | |
| """Fetch call recording by UniqueId and save to /tmp. Returns (filepath, url).""" | |
| url = f"{BASE_URL}/calllogs/{CLIENT_ID}/{unique_id}" | |
| r = requests.get(url, headers=AUTH_HEADERS, timeout=120) | |
| r.raise_for_status() | |
| path = f"/tmp/call_{unique_id}.mp3" | |
| with open(path, "wb") as f: | |
| f.write(r.content) | |
| return path, url | |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ | |
| # Prompt templates & model options | |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ | |
| PROMPT_TEMPLATES = { | |
| "simple": ( | |
| "You are a call-center conversation analyst for a medical clinic. From the call recording, provide a brief summary:\n" | |
| "- Purpose of the call (appointment / results / complaint / billing / other).\n" | |
| "- Patient intent and expectations.\n" | |
| "- Outcome (booked / call-back / routed / unresolved).\n" | |
| "- Next steps (owner and when).\n" | |
| "- Patient emotion (1β5) and agent tone (1β5).\n" | |
| "- Alerts: urgency/risks/privacy.\n\n" | |
| "Keep it short (6β8 lines). End with a line: βService quality rating: X/5β and one sentence explaining the rating." | |
| ), | |
| "medium": ( | |
| "Act as a senior service analyst. Analyze the call using this structure:\n" | |
| "1) Quick overview: reason for the call, intent, key facts, urgency (low/medium/high).\n" | |
| "2) Call flow (2β4 bullets): what was asked/answered, where friction occurred.\n" | |
| "3) Outcomes & tasks: concrete next actions for clinic/patient with timeframes.\n" | |
| "4) Emotions & empathy: patient mood; agent empathy (0β5).\n" | |
| "5) Procedural compliance: identity verification, disclosure of recording (if stated), no off-protocol medical advice, data accuracy.\n" | |
| "6) Quality rating (0β100) using rubric: greeting, verification, accuracy, empathy, issue resolution (each 0β20)." | |
| ), | |
| "detailed": ( | |
| "You are a quality & operations analyst. Provide an in-depth analysis:\n" | |
| "A) Segmentation: split the call into stages with approximate timestamps (if available) and roles (Patient/Agent).\n" | |
| "B) Structured data for booking: full name (if stated), date of birth, phone, symptoms/complaints (list), onset/duration, possible pain level 0β10 (if mentioned), required specialist/service, preferred time windows, constraints.\n" | |
| "C) Triage & risks: class (routine/urgent/emergency), red flags, whether immediate escalation is needed.\n" | |
| "D) Compliance audit: identity/privacy checks, recording disclosure, consent to data processing, booking policies.\n" | |
| "E) Conversation metrics: talk ratio (agent/patient), interruptions, long pauses, notable keywords.\n" | |
| "F) Coaching for the agent: 3β5 concrete improvements with sample phrasing.\n\n" | |
| "Deliver: (1) A short patient-chart summary (2β3 sentences). (2) A task table with columns: priority, owner, due." | |
| ), | |
| } | |
| TPL_OPTIONS = [ | |
| ("Simple", "simple"), | |
| ("Medium", "medium"), | |
| ("Detailed", "detailed"), | |
| ("Custom", "custom"), | |
| ] | |
| LANG_OPTIONS = [ | |
| ("Russian", "ru"), | |
| ("Auto", "default"), | |
| ("Belarusian", "be"), | |
| ("English", "en"), | |
| ] | |
| MODEL_OPTIONS = [ | |
| ("flash", "models/gemini-2.5-flash"), | |
| ("pro", "models/gemini-2.5-pro"), | |
| ("flash-lite", "models/gemini-2.5-flash-lite"), | |
| ] | |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ | |
| # Utilities | |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ | |
| def label_row(row: dict) -> str: | |
| start = row.get("Start", "") | |
| src = row.get("CallerId", "") | |
| dst = row.get("Destination", "") | |
| dur = row.get("Duration", "") | |
| return f"{start} | {src} β {dst} ({dur}s)" | |
| def _resolve_model(client: "genai.Client", preferred: str) -> str: | |
| name = preferred if preferred.startswith("models/") else f"models/{preferred}" | |
| try: | |
| models = list(client.models.list()) | |
| desired_short = name.split("/", 1)[1] | |
| for m in models: | |
| mname = getattr(m, "name", "") | |
| short = mname.split("/", 1)[1] if mname.startswith("models/") else mname | |
| methods = set(getattr(m, "supported_generation_methods", []) or []) | |
| if short == desired_short and ("generateContent" in methods or not methods): | |
| return f"models/{short}" | |
| # Fallback to first available | |
| for title, candidate in MODEL_OPTIONS: | |
| try: | |
| short = candidate.split("/", 1)[1] | |
| for m in models: | |
| mname = getattr(m, "name", "") | |
| sm = mname.split("/", 1)[1] if mname.startswith("models/") else mname | |
| methods = set(getattr(m, "supported_generation_methods", []) or []) | |
| if sm == short and ("generateContent" in methods or not methods): | |
| return candidate | |
| except Exception: | |
| pass | |
| except Exception: | |
| pass | |
| return name | |
| def _system_instruction(lang_code: str) -> str: | |
| if lang_code == "be": | |
| return "Reply in Belarusian." | |
| if lang_code == "ru": | |
| return "Reply in Russian." | |
| if lang_code == "en": | |
| return "Reply in English." | |
| return "Reply in the caller's language; if unclear, use concise professional English." | |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ | |
| # Gradio handlers | |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ | |
| def ui_fetch_calls(date_str: str): | |
| try: | |
| items = fetch_calllogs(date_str.strip()) | |
| df = pd.DataFrame(items) | |
| opts = [(label_row(r), i) for i, r in df.iterrows()] | |
| msg = f"Calls found: {len(df)}" | |
| # Update dropdown choices and default value | |
| dd = gr.update(choices=[(lbl, idx) for lbl, idx in opts], value=(opts[0][1] if opts else None)) | |
| return df, dd, msg | |
| except requests.HTTPError as e: | |
| body = "" | |
| try: | |
| body = e.response.text[:800] | |
| except Exception: | |
| pass | |
| return pd.DataFrame(), gr.update(choices=[], value=None), f"HTTP error: {e}\n{body}" | |
| except Exception as e: | |
| return pd.DataFrame(), gr.update(choices=[], value=None), f"Load error: {e}" | |
| def ui_play_audio(selected_idx: Optional[int], df: pd.DataFrame): | |
| if selected_idx is None or df is None or df.empty: | |
| return "<em>First fetch the list and select a row.</em>", None, None, "" | |
| try: | |
| row = df.iloc[int(selected_idx)] | |
| except Exception: | |
| return "<em>Invalid row selection.</em>", None, None, "" | |
| unique_id = str(row.get("UniqueId")) | |
| try: | |
| fpath = f"/tmp/call_{unique_id}.mp3" | |
| url_used = f"{BASE_URL}/calllogs/{CLIENT_ID}/{unique_id}" | |
| # Download only if not exists (avoid re-fetch) | |
| if not os.path.exists(fpath) or os.path.getsize(fpath) == 0: | |
| fpath, url_used = fetch_mp3_by_unique_id(unique_id) | |
| html = f'URL: <a href="{url_used}" target="_blank">{url_used}</a>' | |
| return html, fpath, fpath, "Ready β " | |
| except requests.HTTPError as e: | |
| body = "" | |
| try: | |
| body = e.response.text[:800] | |
| except Exception: | |
| pass | |
| return f"HTTP error: {e}<br><pre>{body}</pre>", None, None, "" | |
| except Exception as e: | |
| return f"Playback failed: {e}", None, None, "" | |
| def ui_toggle_custom_prompt(template_key: str): | |
| return gr.update(visible=(template_key == "custom")) | |
| def ui_analyze(selected_idx: Optional[int], df: pd.DataFrame, | |
| template_key: str, custom_prompt: str, lang_code: str, model_pref: str): | |
| if df is None or df.empty or selected_idx is None: | |
| return "First fetch the list, choose a call, and (optionally) click βπ§ Playβ." | |
| if not _HAS_GENAI: | |
| return "β google-genai library not found. Make sure it's in requirements.txt." | |
| try: | |
| row = df.iloc[int(selected_idx)] | |
| except Exception: | |
| return "Invalid row selection." | |
| unique_id = str(row.get("UniqueId")) | |
| mp3_path = f"/tmp/call_{unique_id}.mp3" | |
| # Ensure audio file exists (download if needed) | |
| try: | |
| if not os.path.exists(mp3_path) or os.path.getsize(mp3_path) == 0: | |
| mp3_path, _ = fetch_mp3_by_unique_id(unique_id) | |
| except Exception as e: | |
| return f"Failed to obtain audio for analysis: {e}" | |
| api_key = os.environ.get("GOOGLE_API_KEY", "").strip() | |
| if not api_key: | |
| return "GOOGLE_API_KEY is not set in Space Secrets. Add it in Settings β Secrets and restart the Space." | |
| try: | |
| client = genai.Client(api_key=api_key) | |
| except Exception as e: | |
| return f"Failed to initialize the client: {e}" | |
| # Upload file | |
| try: | |
| uploaded_file = client.files.upload(file=mp3_path) | |
| except Exception as e: | |
| return f"File upload error: {e}" | |
| # Prepare prompt | |
| if template_key == "custom": | |
| prompt = (custom_prompt or "").strip() or PROMPT_TEMPLATES["simple"] | |
| else: | |
| prompt = PROMPT_TEMPLATES.get(template_key, PROMPT_TEMPLATES["simple"]) | |
| sys_inst = _system_instruction(lang_code) | |
| model_name = _resolve_model(client, model_pref) | |
| # Call model | |
| try: | |
| merged = f"""[SYSTEM INSTRUCTION: {sys_inst}] | |
| {prompt}""" | |
| resp = client.models.generate_content(model=model_name, contents=[uploaded_file, merged]) | |
| text = getattr(resp, "text", None) | |
| if not text: | |
| return "Analysis finished but returned no text. Check model settings and file format." | |
| return f"### Analysis result\n\n{text}" | |
| except Exception as e: | |
| # Try to attach more error details | |
| msg = str(e) | |
| try: | |
| if hasattr(e, "args") and e.args: | |
| msg = msg + "\n\n" + str(e.args[0]) | |
| except Exception: | |
| pass | |
| return f"Error during model call: {msg}" | |
| finally: | |
| # Best-effort cleanup of remote file | |
| try: | |
| if 'uploaded_file' in locals() and hasattr(uploaded_file, 'name'): | |
| client.files.delete(name=uploaded_file.name) | |
| except Exception: | |
| pass | |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ | |
| # Build Gradio UI | |
| # βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ | |
| def _today_str(): | |
| return _dt.date.today().strftime("%Y-%m-%d") | |
| with gr.Blocks(title="Vochi CRM Call Logs (Gradio)") as demo: | |
| gr.Markdown( | |
| """ | |
| # Vochi CRM β MP3 β AI analysis | |
| *Fetch daily calls, play/download MP3, and analyze the call with an AI model.* | |
| """ | |
| ) | |
| with gr.Tabs(): | |
| with gr.Tab("Vochi CRM"): | |
| with gr.Row(): | |
| date_inp = gr.Textbox(label="Date", value=_today_str(), scale=1) | |
| fetch_btn = gr.Button("Fetch list", variant="primary", scale=0) | |
| calls_df = gr.Dataframe(value=pd.DataFrame(), label="Call list", interactive=False) | |
| row_dd = gr.Dropdown(choices=[], label="Call", info="Select a row for playback/analysis") | |
| with gr.Row(): | |
| play_btn = gr.Button("π§ Play") | |
| url_html = gr.HTML() | |
| audio_out = gr.Audio(label="Audio", type="filepath") | |
| file_out = gr.File(label="MP3 download") | |
| status_fetch = gr.Markdown() | |
| with gr.Tab("AI Analysis"): | |
| with gr.Row(): | |
| tpl_dd = gr.Dropdown(choices=TPL_OPTIONS, value="simple", label="Template") | |
| lang_dd = gr.Dropdown(choices=LANG_OPTIONS, value="default", label="Language") | |
| model_dd = gr.Dropdown(choices=MODEL_OPTIONS, value="models/gemini-2.5-flash", label="Model") | |
| custom_prompt_tb = gr.Textbox(label="Custom prompt", lines=8, visible=False) | |
| analyze_btn = gr.Button("π§ Analyze", variant="primary") | |
| analysis_md = gr.Markdown() | |
| # Wire events | |
| fetch_btn.click(ui_fetch_calls, inputs=[date_inp], outputs=[calls_df, row_dd, status_fetch]) | |
| play_btn.click(ui_play_audio, inputs=[row_dd, calls_df], outputs=[url_html, audio_out, file_out, status_fetch]) | |
| tpl_dd.change(ui_toggle_custom_prompt, inputs=[tpl_dd], outputs=[custom_prompt_tb]) | |
| analyze_btn.click( | |
| ui_analyze, | |
| inputs=[row_dd, calls_df, tpl_dd, custom_prompt_tb, lang_dd, model_dd], | |
| outputs=[analysis_md], | |
| ) | |
| if __name__ == "__main__": | |
| # On HF Spaces, just running this file is enough; launch() is fine for local dev, too. | |
| demo.launch() | |