Spaces:
Sleeping
Sleeping
| from transformers import AutoTokenizer | |
| model_nm = 't5-base' | |
| device = 'cuda' | |
| def tokenize_for_inference(text): | |
| tokenizer = AutoTokenizer.from_pretrained(model_nm) | |
| model_inputs = tokenizer.encode( | |
| text, | |
| max_length = 512, | |
| padding=True, | |
| truncation=True, | |
| return_tensors='pt' | |
| ) | |
| return model_inputs.to(device) | |