optimum-cli export onnx \
--model google/gemma-2b-it \
--opset 17 \
--framework pt \
~/.onnx/gemma-2b-it-pt
c.f. https://github.com/huggingface/optimum/issues/1755
fail
python -m onnxruntime_genai.models.builder \
-m ~/.onnx/gemma-2b-it-pt \
-o ~/.onnx/gemma-2b-it-onnx-int4-cpu \
-p int4 \
-e cpu
python \
-m onnxruntime_genai.models.builder \
-m google/gemma-2-2b-it \
-p fp16 \
-e cpu \
-o ~/.onnx/gemma-2b-it-onnx-fp16-cpu
- Downloads last month
- 11
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support