File size: 790 Bytes
ed8a7fb | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 | from optimum.onnxruntime import ORTModelForSequenceClassification
from transformers import AutoTokenizer
import torch
class EndpointHandler():
def __init__(self, path=""):
# load the optimized model
self.model = ORTModelForSequenceClassification.from_pretrained(path)
self.tokenizer = AutoTokenizer.from_pretrained(path)
def __call__(self, data):
answers = data.pop("answers")
paraphrases = data.pop("paraphrases")
inputs = self.tokenizer(answers, paraphrases, max_length=253, padding=True, truncation=True, return_tensors='pt')
with torch.no_grad():
outputs = self.model(**inputs)
logits = outputs.logits
predictions = torch.argmax(logits, dim=-1).numpy()
return list(predictions) |