code

#1
by asdgad - opened

from transformers import AutoTokenizer, pipeline

from optimum.onnxruntime import ORTModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("optimum/mistral-1.1b-testing")

model = ORTModelForCausalLM.from_pretrained("optimum/mistral-1.1b-testing")

#model = ORTModelForCausalLM.from_pretrained("/content/1")

onnx_gen = pipeline("text-generation", model=model, tokenizer=tokenizer)

text = "My name is Arthur and I live in"

gen = onnx_gen(text)
print(gen)

!pip install "optimum-onnx[onnxruntime]"

Sign up or log in to comment