Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -2,6 +2,7 @@ from threading import Thread
|
|
| 2 |
import gradio as gr
|
| 3 |
import random
|
| 4 |
import torch
|
|
|
|
| 5 |
from transformers import (
|
| 6 |
AutoModelForCausalLM,
|
| 7 |
AutoTokenizer,
|
|
@@ -10,8 +11,8 @@ from transformers import (
|
|
| 10 |
)
|
| 11 |
|
| 12 |
# Constants for the model and configuration
|
| 13 |
-
MODEL_ID = "
|
| 14 |
-
WINDOW_SIZE =
|
| 15 |
DEVICE = "cuda"
|
| 16 |
|
| 17 |
# Load model configuration, tokenizer, and model
|
|
@@ -35,7 +36,6 @@ GREETING_MESSAGES = [
|
|
| 35 |
"The universe awaits! I'm AstroSage. What astronomical wonders shall we discuss?",
|
| 36 |
]
|
| 37 |
|
| 38 |
-
|
| 39 |
def generate_text(prompt: str, history: list, max_new_tokens=512, temperature=0.7, top_p=0.95, top_k=50):
|
| 40 |
"""
|
| 41 |
Generate a response using the transformer model.
|
|
@@ -86,7 +86,7 @@ def user(user_message, history):
|
|
| 86 |
history = []
|
| 87 |
return "", history + [{"role": "user", "content": user_message}]
|
| 88 |
|
| 89 |
-
|
| 90 |
def bot(history):
|
| 91 |
"""
|
| 92 |
Generate the bot's response based on the history.
|
|
|
|
| 2 |
import gradio as gr
|
| 3 |
import random
|
| 4 |
import torch
|
| 5 |
+
import spaces
|
| 6 |
from transformers import (
|
| 7 |
AutoModelForCausalLM,
|
| 8 |
AutoTokenizer,
|
|
|
|
| 11 |
)
|
| 12 |
|
| 13 |
# Constants for the model and configuration
|
| 14 |
+
MODEL_ID = "AstroMLab/AstroSage-8B"
|
| 15 |
+
WINDOW_SIZE = 2048
|
| 16 |
DEVICE = "cuda"
|
| 17 |
|
| 18 |
# Load model configuration, tokenizer, and model
|
|
|
|
| 36 |
"The universe awaits! I'm AstroSage. What astronomical wonders shall we discuss?",
|
| 37 |
]
|
| 38 |
|
|
|
|
| 39 |
def generate_text(prompt: str, history: list, max_new_tokens=512, temperature=0.7, top_p=0.95, top_k=50):
|
| 40 |
"""
|
| 41 |
Generate a response using the transformer model.
|
|
|
|
| 86 |
history = []
|
| 87 |
return "", history + [{"role": "user", "content": user_message}]
|
| 88 |
|
| 89 |
+
@spaces.GPU
|
| 90 |
def bot(history):
|
| 91 |
"""
|
| 92 |
Generate the bot's response based on the history.
|