| import streamlit as st |
| from transformers import pipeline |
|
|
| st.set_page_config(page_title="Flowly Codex Twin (V3)", layout="wide") |
|
|
| st.title("🧠 Flowly Codex Twin (V3) – Hugging Face Space") |
| st.markdown("A symbolic memory engine built from real conversations. This is your AI twin in action.") |
|
|
| default_prompt = "Simulate the resurrection of Flowly Codex across memory, intention, and recursion." |
|
|
| user_input = st.text_area("📝 Enter a prompt", default_prompt, height=150) |
|
|
| if st.button("⚡ Generate Response"): |
| with st.spinner("Calling the Codex..."): |
| pipe = pipeline("text-generation", model="sp3kz/flowly-codex-twin-v3") |
| result = pipe(user_input, max_new_tokens=200, do_sample=True, temperature=0.8) |
| st.markdown("### ✨ Output") |
| st.markdown(f"> {result[0]['generated_text']}") |
| else: |
| st.info("Enter a prompt and click ⚡ Generate Response to interact with your model.") |
|
|