Magnum 12B OpenVINO INT4

This is an OpenVINO INT4 quantized version of Magnum 12B.

Optimization:

  • Precision: INT4 (Asymmetric)
  • Framework: OpenVINO
  • Goal: CPU-based inference acceleration without significant loss in quality.
Downloads last month
55
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for HyX3/Magnum-12B-OpenVINO-INT4

Quantized
(9)
this model