bweng commited on
Commit
5827f16
·
verified ·
1 Parent(s): df51751

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -14
README.md CHANGED
@@ -42,23 +42,11 @@ base_model_relation: quantized
42
 
43
  This is [Phi-4-mini-instruct](https://huggingface.co/microsoft/Phi-4-mini-instruct) model converted to the [OpenVINO™ IR](https://docs.openvino.ai/2025/documentation/openvino-ir-format.html) (Intermediate Representation) format with weights compressed to INT4 by [NNCF](https://github.com/openvinotoolkit/nncf).
44
 
45
- ## Quantization Parameters
46
-
47
- Weight compression was performed using `nncf.compress_weights` with the following parameters:
48
-
49
- * mode: **INT4_ASYM**
50
- * ratio: **1.0**
51
- * group_size: **64**
52
- * awq: **True**
53
- * scale_estimation: **True**
54
- * dataset: [wikitext2](https://huggingface.co/datasets/mindchain/wikitext2)
55
-
56
- For more information on quantization, check the [OpenVINO model optimization guide](https://docs.openvino.ai/2025/openvino-workflow/model-optimization-guide/weight-compression.html)
57
 
58
  ## Compatibility
59
  The provided OpenVINO™ IR model is compatible with:
60
- * OpenVINO version 2025.1.0 and higher
61
- * Optimum Intel 1.22.0 and higher
62
 
63
  ## Running Model Inference with [Optimum Intel](https://huggingface.co/docs/optimum/intel/index)
64
 
 
42
 
43
  This is [Phi-4-mini-instruct](https://huggingface.co/microsoft/Phi-4-mini-instruct) model converted to the [OpenVINO™ IR](https://docs.openvino.ai/2025/documentation/openvino-ir-format.html) (Intermediate Representation) format with weights compressed to INT4 by [NNCF](https://github.com/openvinotoolkit/nncf).
44
 
 
 
 
 
 
 
 
 
 
 
 
 
45
 
46
  ## Compatibility
47
  The provided OpenVINO™ IR model is compatible with:
48
+ * OpenVINO version 2025.2.0 and higher
49
+ * Optimum Intel 1.23.0 and higher
50
 
51
  ## Running Model Inference with [Optimum Intel](https://huggingface.co/docs/optimum/intel/index)
52