from transformers import AutoTokenizer
from optimum.onnxruntime import ORTModelForQuestionAnswering
import optimum.onnxruntime

tokenizer = AutoTokenizer.from_pretrained("typeof/distilbert-base-cased-finetuned-conll03-english-quantized")
model = optimum.onnxruntime.ORTModelForTokenClassification.from_pretrained("typeof/distilbert-base-cased-finetuned-conll03-english-quantized")
Downloads last month
4
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support