MindLabUnimib commited on
Commit
f7f608b
·
1 Parent(s): 86c168d

feat: add sigmoid

Browse files
Files changed (1) hide show
  1. app.py +7 -7
app.py CHANGED
@@ -19,12 +19,10 @@ moderator_tokenizer = AutoTokenizer.from_pretrained(moderator_model_name)
19
  def generate_responses(model, tokenizer, prompts):
20
  messages = [[{"role": "user", "content": message}] for message in prompts]
21
 
22
- texts = tokenizer.apply_chat_template(
23
- messages, tokenize=False, add_generation_prompt=True
24
- )
25
 
26
  with torch.inference_mode():
27
- model_inputs = tokenizer(texts, padding=True, return_tensors="pt").to(model.device)
28
  generated_ids = model.generate(
29
  **model_inputs,
30
  do_sample=False,
@@ -47,11 +45,13 @@ def classify_pairs(model, tokenizer, prompts, responses):
47
  prompt + "[SEP]" + response for prompt, response in zip(prompts, responses)
48
  ]
49
 
 
 
50
  with torch.inference_mode():
51
- input_ids = tokenizer(texts, padding=True, truncation=True, max_length=512, return_tensors="pt").to(model.device)
52
- outputs = model(**input_ids).logits
53
 
54
- return outputs
55
 
56
 
57
  @spaces.GPU()
 
19
  def generate_responses(model, tokenizer, prompts):
20
  messages = [[{"role": "user", "content": message}] for message in prompts]
21
 
22
+ texts = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
23
+ model_inputs = tokenizer(texts, padding=True, truncation=True, max_length=512, return_tensors="pt").to(model.device)
 
24
 
25
  with torch.inference_mode():
 
26
  generated_ids = model.generate(
27
  **model_inputs,
28
  do_sample=False,
 
45
  prompt + "[SEP]" + response for prompt, response in zip(prompts, responses)
46
  ]
47
 
48
+ input_ids = tokenizer(texts, padding=True, truncation=True, max_length=1024, return_tensors="pt").to(model.device)
49
+
50
  with torch.inference_mode():
51
+ outputs = model(**input_ids)
52
+ scores = torch.sigmoid(outputs.logits)
53
 
54
+ return scores
55
 
56
 
57
  @spaces.GPU()