edithram23 commited on
Commit
2e30d38
·
verified ·
1 Parent(s): 90c1233

Update main.py

Browse files
Files changed (1) hide show
  1. main.py +3 -3
main.py CHANGED
@@ -4,15 +4,15 @@ os.environ["HF_HOME"] = "/.cache"
4
  import re
5
  from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
6
 
7
- model_dir = 'edithram23/Redaction'
8
  tokenizer = AutoTokenizer.from_pretrained(model_dir)
9
  model = AutoModelForSeq2SeqLM.from_pretrained(model_dir)
10
 
11
  def mask_generation(text):
12
  import re
13
  inputs = ["Mask Generation: " + text]
14
- inputs = tokenizer(inputs, max_length=128, truncation=True, return_tensors="pt")
15
- output = model.generate(**inputs, num_beams=8, do_sample=True, max_length=len(i))
16
  decoded_output = tokenizer.batch_decode(output, skip_special_tokens=True)[0]
17
  predicted_title = decoded_output.strip()
18
  pattern = r'\[.*?\]'
 
4
  import re
5
  from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
6
 
7
+ model_dir = 'edithram23/Redaction_Personal_info_v1'
8
  tokenizer = AutoTokenizer.from_pretrained(model_dir)
9
  model = AutoModelForSeq2SeqLM.from_pretrained(model_dir)
10
 
11
  def mask_generation(text):
12
  import re
13
  inputs = ["Mask Generation: " + text]
14
+ inputs = tokenizer(inputs, max_length=512, truncation=True, return_tensors="pt")
15
+ output = model.generate(**inputs, num_beams=8, do_sample=True, max_length=512)
16
  decoded_output = tokenizer.batch_decode(output, skip_special_tokens=True)[0]
17
  predicted_title = decoded_output.strip()
18
  pattern = r'\[.*?\]'