--- library_name: transformers tags: [] --- # USAGE ```python import torch from transformers import AutoModelForSequenceClassification, AutoTokenizer MODEL_NAME = "swarogthehater/IMAGE_INTENT" model = AutoModelForSequenceClassification.from_pretrained(MODEL_NAME, load_in_8bit=True) model.eval() tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME) text = "show me yourself" IMAGE_INTENT = "[IMG]" input_text = text+"[SEP]"+IMAGE_INTENT+"[SEP]" device = torch.device("cpu") batch = tokenizer.encode_plus(input_text, return_tensors="pt") input_ids = batch['input_ids'].to(device) attention_mask = batch['attention_mask'].to(device) token_type_ids = batch['token_type_ids'].to(device) outputs = model(input_ids, attention_mask=attention_mask, token_type_ids=token_type_ids) #labels in common nli terms (0: entailment, 1: neutral, 2: contradiction) print(outputs.logits.argmax().item()) #labels for img intent label = 1 if outputs.logits.argmax().item() == 0 else 0 print(label) #scores print(outputs.logits.float().softmax(dim=-1).detach().numpy()) ```