Spaces:
Running
Running
File size: 7,552 Bytes
ccee4a3 81c1320 836f78e 489ca1d 685d841 489ca1d fce46f1 489ca1d 51fde0f 489ca1d e317b88 489ca1d fce46f1 175538c 489ca1d 5f8596a 836f78e fce46f1 489ca1d d824a95 0c10f52 fce46f1 489ca1d ce44f5d 489ca1d 8351828 489ca1d 81c1320 489ca1d 8351828 489ca1d 8351828 489ca1d 3ddb7ac 489ca1d 3ddb7ac 8351828 489ca1d 101b43d 8351828 ce44f5d 489ca1d e0fa056 8351828 489ca1d 2e8bbbb 489ca1d 8351828 489ca1d 8351828 5f8596a | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 | import gradio as gr
from huggingface_hub import InferenceClient
from array import array
from functools import lru_cache
import os, re, time
# 1. API Configuration - Locked to the stable 7B model
HF_TOKEN = os.getenv("HF_TOKEN")
MODEL_ID = "Qwen/Qwen2.5-7B-Instruct"
client = InferenceClient(MODEL_ID, token=HF_TOKEN)
# 2. T3 High-Speed Logic Kernel
class StateController:
__slots__ = ("_state", "_rom60", "_symbols", "_rendered")
def __init__(self):
self._state = array("B", [0]) * 121
self._rom60 = tuple(tuple((i * j) % 60 for j in range(60)) for i in range(60))
self._symbols = "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz01234567"
self._rendered = "".join(" [NODE_120] " if i == 120 else ("<" if i % 10 == 0 else ".") for i in range(121))
@lru_cache(maxsize=128)
def compute_distribution(self, total, nodes) -> str:
if nodes <= 0: return "Error: Node count must be positive."
base, rem = divmod(total, nodes)
res = f"T3 Logic Kernel resolved {total} units across {nodes} nodes:\n\n"
for i in range(nodes):
res += f"NODE_{i+1:02}: {base + (1 if i < rem else 0)} units\n"
return res
def get_glyphs(self) -> str:
return f"Rendering 121-point state array:\n\n{self._rendered}\n\nSystem State: RESOLVED"
def generate_receipt(self, a, b, c) -> str:
idx = (self._rom60[a % 60][b % 60] ^ (c % 60)) % 60
return f"0{self._symbols[idx]}"
def validate_receipt(self, receipt, a, b, c) -> str:
expected = self.generate_receipt(a, b, c)
if receipt == expected:
return f"√ CHECKSUM VALID: Receipt {receipt} verified for allocation ({a}, {b}, {c})."
return f"× CHECKSUM INVALID: Expected {expected}, received {receipt}."
controller = StateController()
def format_telemetry(seconds: float) -> str:
if seconds < 0.001: return f"{seconds * 1_000_000:.2f} \u03BCs"
return f"{seconds * 1_000:.2f} ms" if seconds < 1 else f"{seconds:.2f} s"
# 3. Core Response Logic
def generate_responses(user_message, p_hist, c_hist):
msg = user_message.strip()
if not msg: yield p_hist or [], c_hist or [], ""; return
p_hist, c_hist = p_hist or [], c_hist or []
p_hist.append({"role": "user", "content": msg})
p_hist.append({"role": "assistant", "content": ""})
c_hist.append({"role": "user", "content": msg})
c_hist.append({"role": "assistant", "content": ""})
yield p_hist, c_hist, ""
start_time = time.perf_counter()
# --- LOCAL INTERCEPTORS ---
dist_match = re.search(r"(\d+)\s+units\s+across\s+(\d+)\s+nodes", msg, re.IGNORECASE)
diag_match = any(kw in msg.lower() for kw in ["diagnostic", "grid"])
rcpt_match = re.search(r"verify receipt\s+([a-zA-Z0-9]{2})\s+for\s+(\d+),\s*(\d+),\s*(\d+)", msg, re.IGNORECASE)
if dist_match or diag_match or rcpt_match:
if dist_match:
res = controller.compute_distribution(int(dist_match.group(1)), int(dist_match.group(2)))
elif rcpt_match:
res = controller.validate_receipt(rcpt_match.group(1), int(rcpt_match.group(2)), int(rcpt_match.group(3)), int(rcpt_match.group(4)))
else:
res = controller.get_glyphs()
elapsed = time.perf_counter() - start_time
p_hist[-1]["content"] = f"{res}\n\n---\n*Telemetry: {format_telemetry(elapsed)} | Source: LOCAL T3 KERNEL*"
yield p_hist, c_hist, ""
else:
try:
res_text = ""
stream = client.chat_completion(
messages=[{"role":"system","content":"T3 Augmented Logic Engine"}] + p_hist[:-1],
max_tokens=512, stream=True, temperature=0.1
)
for chunk in stream:
res_text += (chunk.choices[0].delta.content or "")
p_hist[-1]["content"] = res_text
yield p_hist, c_hist, ""
p_hist[-1]["content"] += f"\n\n---\n*Telemetry: {format_telemetry(time.perf_counter()-start_time)} | Source: AUGMENTED CLOUD*"
yield p_hist, c_hist, ""
except Exception as e:
p_hist[-1]["content"] = f"Primary Error: {str(e)}"
yield p_hist, c_hist, ""
comp_start = time.perf_counter()
c_hist[-1]["content"] = "*Routing through standard infrastructure...*"
yield p_hist, c_hist, ""
try:
res_text = ""
stream = client.chat_completion(
messages=[{"role":"system","content":"Vanilla AI"}] + c_hist[:-1],
max_tokens=512, stream=True, temperature=0.7
)
for chunk in stream:
res_text += (chunk.choices[0].delta.content or "")
c_hist[-1]["content"] = res_text
yield p_hist, c_hist, ""
c_hist[-1]["content"] += f"\n\n---\n*Telemetry: {format_telemetry(time.perf_counter()-comp_start)} | Source: VANILLA CLOUD*"
yield p_hist, c_hist, ""
except Exception as e:
c_hist[-1]["content"] = f"Competitor Error: {str(e)}"
yield p_hist, c_hist, ""
# 4. Interface Build (With Scrollable Container & NO 'type' attributes)
custom_css = """
body, .gradio-container { background-color: #110c08 !important; color: #fb923c !important; }
footer { display: none !important; }
#scrollable-box { max-height: 160px; overflow-y: auto; border: 1px solid #333; padding: 5px; border-radius: 8px; margin-bottom: 10px; }
"""
example_prompts = [
["Run grid diagnostic"],
["Calculate the integer distribution for 50000 units across 12 nodes."],
["Define P vs. NP. Then validate a 120-unit distribution across 3 nodes."],
["Execute a Tier-3 Distribution Audit for 8593 units across 14 nodes."],
["Verify receipt 0e for 60, 30, 30"],
["Distribute 1000000 units across 7 nodes."],
["Perform a hardware grid initialization and diagnostic check."],
["Allocate exactly 2048 units across 16 nodes for cluster balancing."],
["Explain the theory of relativity. Then process 999 units across 9 nodes."],
["Run a full system diagnostic on the logical array."],
["Load balance 123456789 units across 256 nodes."],
["Draft an email to the logistics team. Then route 400 units across 5 nodes."],
["Initialize grid memory matrix and verify logic gate alignment."],
["Evaluate node efficiency for 7777 units across 11 nodes."],
["Explain how standard AI struggles with deterministic mathematical verification."]
]
with gr.Blocks() as demo:
gr.Markdown("# [ GLYPH.IO ]\n### Dual-Engine Hardware Benchmark")
# 100% clean Chatbots with NO 'type="messages"' argument to prevent crashes
p_chat = gr.Chatbot(label="Augmented Logic Kernel (T3 Architecture)", height=350)
with gr.Row():
msg_in = gr.Textbox(label="Message", placeholder="Test P vs NP or Logistics Distribution...", scale=8)
submit_btn = gr.Button("Execute", scale=1, variant="primary")
with gr.Column(elem_id="scrollable-box"):
gr.Examples(examples=example_prompts, inputs=msg_in, label="Diagnostic Test Suite (Scroll for more)")
# 100% clean Chatbots with NO 'type="messages"' argument to prevent crashes
c_chat = gr.Chatbot(label="Vanilla Qwen 2.5 (Standard Infrastructure)", height=350)
msg_in.submit(generate_responses, [msg_in, p_chat, c_chat], [p_chat, c_chat, msg_in])
submit_btn.click(generate_responses, [msg_in, p_chat, c_chat], [p_chat, c_chat, msg_in])
if __name__ == "__main__":
demo.queue().launch(theme=gr.themes.Soft(primary_hue="orange"), css=custom_css)
|