Spaces:
Sleeping
Sleeping
Keira James
commited on
Commit
·
80e7abb
1
Parent(s):
7aea9b8
update app
Browse files
app.py
CHANGED
|
@@ -1,6 +1,12 @@
|
|
| 1 |
import streamlit as st
|
|
|
|
| 2 |
import torch
|
| 3 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 4 |
# Function to generate a response
|
| 5 |
def generate_response(prompt):
|
| 6 |
if not prompt:
|
|
|
|
| 1 |
import streamlit as st
|
| 2 |
+
from transformers import AutoTokenizer, AutoModelForCausalLM
|
| 3 |
import torch
|
| 4 |
|
| 5 |
+
# Load the tokenizer and model (change 'model_name' to your specific model)
|
| 6 |
+
model_name = "gpt2" # Replace with your model
|
| 7 |
+
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
| 8 |
+
model = AutoModelForCausalLM.from_pretrained(model_name)
|
| 9 |
+
|
| 10 |
# Function to generate a response
|
| 11 |
def generate_response(prompt):
|
| 12 |
if not prompt:
|