Spaces:
Sleeping
Sleeping
Update main.py
Browse files
main.py
CHANGED
|
@@ -4,15 +4,15 @@ os.environ["HF_HOME"] = "/.cache"
|
|
| 4 |
import re
|
| 5 |
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
|
| 6 |
|
| 7 |
-
model_dir = 'edithram23/
|
| 8 |
tokenizer = AutoTokenizer.from_pretrained(model_dir)
|
| 9 |
model = AutoModelForSeq2SeqLM.from_pretrained(model_dir)
|
| 10 |
|
| 11 |
def mask_generation(text):
|
| 12 |
import re
|
| 13 |
inputs = ["Mask Generation: " + text]
|
| 14 |
-
inputs = tokenizer(inputs, max_length=
|
| 15 |
-
output = model.generate(**inputs, num_beams=8, do_sample=True, max_length=
|
| 16 |
decoded_output = tokenizer.batch_decode(output, skip_special_tokens=True)[0]
|
| 17 |
predicted_title = decoded_output.strip()
|
| 18 |
pattern = r'\[.*?\]'
|
|
|
|
| 4 |
import re
|
| 5 |
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
|
| 6 |
|
| 7 |
+
model_dir = 'edithram23/Redaction_Personal_info_v1'
|
| 8 |
tokenizer = AutoTokenizer.from_pretrained(model_dir)
|
| 9 |
model = AutoModelForSeq2SeqLM.from_pretrained(model_dir)
|
| 10 |
|
| 11 |
def mask_generation(text):
|
| 12 |
import re
|
| 13 |
inputs = ["Mask Generation: " + text]
|
| 14 |
+
inputs = tokenizer(inputs, max_length=512, truncation=True, return_tensors="pt")
|
| 15 |
+
output = model.generate(**inputs, num_beams=8, do_sample=True, max_length=512)
|
| 16 |
decoded_output = tokenizer.batch_decode(output, skip_special_tokens=True)[0]
|
| 17 |
predicted_title = decoded_output.strip()
|
| 18 |
pattern = r'\[.*?\]'
|