Update app.py
Browse files
app.py
CHANGED
|
@@ -3,13 +3,18 @@ from transformers import pipeline
|
|
| 3 |
from arabert.aragpt2.grover.modeling_gpt2 import GPT2LMHeadModel
|
| 4 |
from transformers import AutoTokenizer
|
| 5 |
import re
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 6 |
|
| 7 |
model_name = "Naseej/AskMe-Large"
|
| 8 |
tokenizer = AutoTokenizer.from_pretrained(model_name, bos_token='<|startoftext|>',
|
| 9 |
eos_token='<|endoftext|>', pad_token='<|pad|>')
|
| 10 |
-
model = GPT2LMHeadModel.from_pretrained(model_name)
|
| 11 |
model.resize_token_embeddings(len(tokenizer))
|
| 12 |
-
generator = pipeline("text-generation", model=model, tokenizer=tokenizer)
|
| 13 |
|
| 14 |
def generate_response(message, history, num_beams=4, temperature=0.99, do_sample=True, top_k=60, top_p=0.9):
|
| 15 |
prompt = f'Prompt: {message}\nAnswer:'
|
|
|
|
| 3 |
from arabert.aragpt2.grover.modeling_gpt2 import GPT2LMHeadModel
|
| 4 |
from transformers import AutoTokenizer
|
| 5 |
import re
|
| 6 |
+
import torch
|
| 7 |
+
|
| 8 |
+
# Check if CUDA is available
|
| 9 |
+
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
| 10 |
+
print(f"Using device: {device}")
|
| 11 |
|
| 12 |
model_name = "Naseej/AskMe-Large"
|
| 13 |
tokenizer = AutoTokenizer.from_pretrained(model_name, bos_token='<|startoftext|>',
|
| 14 |
eos_token='<|endoftext|>', pad_token='<|pad|>')
|
| 15 |
+
model = GPT2LMHeadModel.from_pretrained(model_name).to(device) # Move model to GPU
|
| 16 |
model.resize_token_embeddings(len(tokenizer))
|
| 17 |
+
generator = pipeline("text-generation", model=model, tokenizer=tokenizer, device=0 if torch.cuda.is_available() else -1)
|
| 18 |
|
| 19 |
def generate_response(message, history, num_beams=4, temperature=0.99, do_sample=True, top_k=60, top_p=0.9):
|
| 20 |
prompt = f'Prompt: {message}\nAnswer:'
|