mobarmg commited on
Commit
b768edc
·
verified ·
1 Parent(s): 73aeb88

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +7 -2
app.py CHANGED
@@ -3,13 +3,18 @@ from transformers import pipeline
3
  from arabert.aragpt2.grover.modeling_gpt2 import GPT2LMHeadModel
4
  from transformers import AutoTokenizer
5
  import re
 
 
 
 
 
6
 
7
  model_name = "Naseej/AskMe-Large"
8
  tokenizer = AutoTokenizer.from_pretrained(model_name, bos_token='<|startoftext|>',
9
  eos_token='<|endoftext|>', pad_token='<|pad|>')
10
- model = GPT2LMHeadModel.from_pretrained(model_name)
11
  model.resize_token_embeddings(len(tokenizer))
12
- generator = pipeline("text-generation", model=model, tokenizer=tokenizer)
13
 
14
  def generate_response(message, history, num_beams=4, temperature=0.99, do_sample=True, top_k=60, top_p=0.9):
15
  prompt = f'Prompt: {message}\nAnswer:'
 
3
  from arabert.aragpt2.grover.modeling_gpt2 import GPT2LMHeadModel
4
  from transformers import AutoTokenizer
5
  import re
6
+ import torch
7
+
8
+ # Check if CUDA is available
9
+ device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
10
+ print(f"Using device: {device}")
11
 
12
  model_name = "Naseej/AskMe-Large"
13
  tokenizer = AutoTokenizer.from_pretrained(model_name, bos_token='<|startoftext|>',
14
  eos_token='<|endoftext|>', pad_token='<|pad|>')
15
+ model = GPT2LMHeadModel.from_pretrained(model_name).to(device) # Move model to GPU
16
  model.resize_token_embeddings(len(tokenizer))
17
+ generator = pipeline("text-generation", model=model, tokenizer=tokenizer, device=0 if torch.cuda.is_available() else -1)
18
 
19
  def generate_response(message, history, num_beams=4, temperature=0.99, do_sample=True, top_k=60, top_p=0.9):
20
  prompt = f'Prompt: {message}\nAnswer:'