Created using Olive-AI: olive optimize --model_name_or_path nvidia/Riva-Translate-4B-Instruct --output_path models/riva_onnx --precision int4 --block_size 64

Downloads last month
-
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for TrentB/Riva-Translate-4B-Instruct-ONNX

Quantized
(8)
this model