TiberiuCristianLeon commited on
Commit
3374d66
·
verified ·
1 Parent(s): 8de4060

inference_mode versus no_grad

Browse files
Files changed (1) hide show
  1. app.py +1 -3
app.py CHANGED
@@ -69,7 +69,6 @@ class Translators:
69
  # model = AutoModel.from_pretrained(self.model_name, trust_remote_code=True)
70
  # model.half() # recommended for GPU
71
  model.eval()
72
- # model.float()
73
  # Translating from one or several sentences to a sole language
74
  src_tokens = tokenizer.encode_source_tokens_to_input_ids(self.input_text, target_language=self.tl)
75
  # src_tokens may be a torch.Tensor or dict depending on tokenizer; ensure it's a tensor
@@ -79,14 +78,13 @@ class Translators:
79
  # # if tokenizer returns dict-like inputs (input_ids, attention_mask)
80
  # for k, v in src_tokens.items():
81
  # src_tokens[k] = v.to(self.device)
82
- # src_tokens = src_tokens.to(self.device)
83
  # generated_tokens = model.generate(src_tokens)
84
  # return tokenizer.batch_decode(generated_tokens, skip_special_tokens=True)[0]
85
  # Translating from one or several sentences to corresponding languages
86
  # src_tokens = tokenizer.encode_source_tokens_to_input_ids_with_different_tags([english_text, english_text, ], target_languages_list=["de", "zh", ])
87
  # generated_tokens = model.generate(src_tokens.to(self.device))
88
  # results = tokenizer.batch_decode(generated_tokens, skip_special_tokens=True)
89
- with torch.no_grad(): # no_grad inference_mode
90
  generated_tokens = model.generate(src_tokens)
91
  result = tokenizer.batch_decode(generated_tokens, skip_special_tokens=True)[0]
92
  return result
 
69
  # model = AutoModel.from_pretrained(self.model_name, trust_remote_code=True)
70
  # model.half() # recommended for GPU
71
  model.eval()
 
72
  # Translating from one or several sentences to a sole language
73
  src_tokens = tokenizer.encode_source_tokens_to_input_ids(self.input_text, target_language=self.tl)
74
  # src_tokens may be a torch.Tensor or dict depending on tokenizer; ensure it's a tensor
 
78
  # # if tokenizer returns dict-like inputs (input_ids, attention_mask)
79
  # for k, v in src_tokens.items():
80
  # src_tokens[k] = v.to(self.device)
 
81
  # generated_tokens = model.generate(src_tokens)
82
  # return tokenizer.batch_decode(generated_tokens, skip_special_tokens=True)[0]
83
  # Translating from one or several sentences to corresponding languages
84
  # src_tokens = tokenizer.encode_source_tokens_to_input_ids_with_different_tags([english_text, english_text, ], target_languages_list=["de", "zh", ])
85
  # generated_tokens = model.generate(src_tokens.to(self.device))
86
  # results = tokenizer.batch_decode(generated_tokens, skip_special_tokens=True)
87
+ with torch.inference_mode(): # no_grad inference_mode
88
  generated_tokens = model.generate(src_tokens)
89
  result = tokenizer.batch_decode(generated_tokens, skip_special_tokens=True)[0]
90
  return result