Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -4,7 +4,16 @@ from sentence_transformers import SentenceTransformer, util
|
|
| 4 |
import random
|
| 5 |
|
| 6 |
# Load models
|
| 7 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 8 |
similarity_model = SentenceTransformer('all-MiniLM-L6-v2')
|
| 9 |
|
| 10 |
# Tone modifier prompts
|
|
@@ -21,7 +30,8 @@ def humanize_text(input_text, tone):
|
|
| 21 |
|
| 22 |
style_prompt = tone_prompts.get(tone, "Paraphrase:")
|
| 23 |
result = paraphraser(f"{style_prompt} {input_text}", max_length=80, num_return_sequences=1)
|
| 24 |
-
output_text =
|
|
|
|
| 25 |
|
| 26 |
# Semantic similarity score
|
| 27 |
emb1 = similarity_model.encode(input_text, convert_to_tensor=True)
|
|
|
|
| 4 |
import random
|
| 5 |
|
| 6 |
# Load models
|
| 7 |
+
from transformers import T5Tokenizer, T5ForConditionalGeneration
|
| 8 |
+
|
| 9 |
+
tokenizer = T5Tokenizer.from_pretrained("Vamsi/T5_Paraphrase_Paws", use_fast=False)
|
| 10 |
+
model = T5ForConditionalGeneration.from_pretrained("Vamsi/T5_Paraphrase_Paws")
|
| 11 |
+
|
| 12 |
+
def generate_paraphrase(text):
|
| 13 |
+
input_ids = tokenizer.encode("paraphrase: " + text, return_tensors="pt", max_length=256, truncation=True)
|
| 14 |
+
output_ids = model.generate(input_ids, max_length=80, num_return_sequences=1, do_sample=True, top_k=120, top_p=0.98)
|
| 15 |
+
return tokenizer.decode(output_ids[0], skip_special_tokens=True)
|
| 16 |
+
|
| 17 |
similarity_model = SentenceTransformer('all-MiniLM-L6-v2')
|
| 18 |
|
| 19 |
# Tone modifier prompts
|
|
|
|
| 30 |
|
| 31 |
style_prompt = tone_prompts.get(tone, "Paraphrase:")
|
| 32 |
result = paraphraser(f"{style_prompt} {input_text}", max_length=80, num_return_sequences=1)
|
| 33 |
+
output_text = generate_paraphrase(input_text)
|
| 34 |
+
|
| 35 |
|
| 36 |
# Semantic similarity score
|
| 37 |
emb1 = similarity_model.encode(input_text, convert_to_tensor=True)
|