Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -85,4 +85,20 @@ with gr.Blocks(title=title) as demo:
|
|
| 85 |
prompt.submit(fn=run, inputs=[prompt, mode, latency], outputs=[output])
|
| 86 |
|
| 87 |
if __name__ == "__main__":
|
| 88 |
-
demo.launch(server_name="0.0.0.0", server_port=7860, share=False)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 85 |
prompt.submit(fn=run, inputs=[prompt, mode, latency], outputs=[output])
|
| 86 |
|
| 87 |
if __name__ == "__main__":
|
| 88 |
+
demo.launch(server_name="0.0.0.0", server_port=7860, share=False)
|
| 89 |
+
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 90 |
+
import gradio as gr
|
| 91 |
+
|
| 92 |
+
|
| 93 |
+
model_id = "x402_hyperlayer_model"
|
| 94 |
+
|
| 95 |
+
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
| 96 |
+
model = AutoModelForCausalLM.from_pretrained(model_id)
|
| 97 |
+
|
| 98 |
+
def chat(prompt):
|
| 99 |
+
inputs = tokenizer(prompt, return_tensors="pt")
|
| 100 |
+
outputs = model.generate(**inputs, max_new_tokens=100)
|
| 101 |
+
return tokenizer.decode(outputs[0], skip_special_tokens=True)
|
| 102 |
+
|
| 103 |
+
demo = gr.Interface(fn=chat, inputs="text", outputs="text", title="X402 HyperLayer Model")
|
| 104 |
+
demo.launch()
|