Spaces:
Sleeping
Sleeping
| from transformers import DistilBertTokenizer | |
| tokenizer = DistilBertTokenizer.from_pretrained("distilbert-base-uncased") | |
| def get_tokens(string): | |
| text = tokenizer(string, return_tensors="pt", padding="max_length", max_length=256, truncation=True) | |
| return { | |
| "input_ids": text["input_ids"], | |
| "attention_mask": text["attention_mask"] | |
| } |