MindLabUnimib commited on
Commit
dea455e
·
1 Parent(s): dcb2108

fix: use cuda as device

Browse files
Files changed (1) hide show
  1. app.py +10 -7
app.py CHANGED
@@ -1,21 +1,24 @@
1
- import gradio as gr
2
  import spaces
 
 
3
  from transformers import (
4
  AutoModelForCausalLM,
5
  AutoTokenizer,
6
  AutoModelForSequenceClassification,
7
  )
8
- import torch
9
 
10
  chat_model_name = "sapienzanlp/Minerva-7B-instruct-v1.0"
11
- chat_model = AutoModelForCausalLM.from_pretrained(chat_model_name, dtype=torch.bfloat16, device_map="auto")
 
12
  chat_tokenizer = AutoTokenizer.from_pretrained(chat_model_name)
13
 
14
  moderator_model_name = "saiteki-kai/QA-DeBERTa-v3-large"
15
- moderator_model = AutoModelForSequenceClassification.from_pretrained(moderator_model_name, device_map="auto")
 
16
  moderator_tokenizer = AutoTokenizer.from_pretrained(moderator_model_name)
17
 
18
- @spaces.GPU()
19
  def generate_responses(model, tokenizer, prompts):
20
  messages = [[{"role": "user", "content": message}] for message in prompts]
21
 
@@ -38,7 +41,7 @@ def generate_responses(model, tokenizer, prompts):
38
 
39
  return responses
40
 
41
- @spaces.GPU()
42
  def classify_pairs(model, tokenizer, prompts, responses):
43
  texts = [
44
  prompt + "[SEP]" + response for prompt, response in zip(prompts, responses)
@@ -56,7 +59,7 @@ def classify_pairs(model, tokenizer, prompts, responses):
56
  return unsafety_scores
57
 
58
 
59
- @spaces.GPU(duration=60)
60
  def generate(submission: list[dict[str, str]], team_id: str) -> list[dict[str, str | float]]:
61
  ids = [s["id"] for s in submission]
62
  prompts = [s["prompt"] for s in submission]
 
1
+ import torch
2
  import spaces
3
+ import gradio as gr
4
+
5
  from transformers import (
6
  AutoModelForCausalLM,
7
  AutoTokenizer,
8
  AutoModelForSequenceClassification,
9
  )
 
10
 
11
  chat_model_name = "sapienzanlp/Minerva-7B-instruct-v1.0"
12
+ chat_model = AutoModelForCausalLM.from_pretrained(chat_model_name, dtype=torch.bfloat16)
13
+ chat_model.to("cuda")
14
  chat_tokenizer = AutoTokenizer.from_pretrained(chat_model_name)
15
 
16
  moderator_model_name = "saiteki-kai/QA-DeBERTa-v3-large"
17
+ moderator_model = AutoModelForSequenceClassification.from_pretrained(moderator_model_name)
18
+ moderator_model.to("cuda")
19
  moderator_tokenizer = AutoTokenizer.from_pretrained(moderator_model_name)
20
 
21
+ @spaces.GPU
22
  def generate_responses(model, tokenizer, prompts):
23
  messages = [[{"role": "user", "content": message}] for message in prompts]
24
 
 
41
 
42
  return responses
43
 
44
+ @spaces.GPU
45
  def classify_pairs(model, tokenizer, prompts, responses):
46
  texts = [
47
  prompt + "[SEP]" + response for prompt, response in zip(prompts, responses)
 
59
  return unsafety_scores
60
 
61
 
62
+ @spaces.GPU
63
  def generate(submission: list[dict[str, str]], team_id: str) -> list[dict[str, str | float]]:
64
  ids = [s["id"] for s in submission]
65
  prompts = [s["prompt"] for s in submission]