cmh commited on
Commit
4d0774a
·
verified ·
1 Parent(s): 1226336

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -1
README.md CHANGED
@@ -25,10 +25,13 @@ pipeline_tag: text-generation
25
  | [phi-4.bf16.q8p](https://huggingface.co/cmh/test/blob/main/phi-4.bf16.q8p.gguf) | 8 bits per weight | 15.6 GB | **18.6 GB** |
26
  | [phi-4.bf16](https://huggingface.co/cmh/test/blob/main/phi-4.bf16.gguf) | 16 bits per weight | 29.3 GB | tbd |
27
 
28
- <sub>*approximate value at **16k context, FP16 cache**.<sub>
 
29
  ---------------------------------------------
 
30
  [ZeroWw quantization: huggingface.co/RobertSinclair](https://huggingface.co/RobertSinclair)
31
 
 
32
  ```
33
  python convert_hf_to_gguf.py --outtype bf16 phi-4 --outfile phi-4.bf16.gguf
34
 
@@ -40,6 +43,7 @@ llama-quantize --imatrix imatrix.dat --leave-output-tensor phi-4.bf16.gguf phi-4
40
  llama-quantize --allow-requantize --output-tensor-type bf16 --token-embedding-type bf16 phi-4.bf16.gguf phi-4.bf16.q8.gguf q8_0
41
  llama-quantize --allow-requantize --pure phi-4.bf16.gguf phi-4.bf16.q8p.gguf q8_0
42
  ```
 
43
  ---------------------------------------------
44
 
45
  # Phi-4 Model Card
 
25
  | [phi-4.bf16.q8p](https://huggingface.co/cmh/test/blob/main/phi-4.bf16.q8p.gguf) | 8 bits per weight | 15.6 GB | **18.6 GB** |
26
  | [phi-4.bf16](https://huggingface.co/cmh/test/blob/main/phi-4.bf16.gguf) | 16 bits per weight | 29.3 GB | tbd |
27
 
28
+ <sub>*approximate value at **16k context, FP16 cache**.<sup>
29
+
30
  ---------------------------------------------
31
+
32
  [ZeroWw quantization: huggingface.co/RobertSinclair](https://huggingface.co/RobertSinclair)
33
 
34
+
35
  ```
36
  python convert_hf_to_gguf.py --outtype bf16 phi-4 --outfile phi-4.bf16.gguf
37
 
 
43
  llama-quantize --allow-requantize --output-tensor-type bf16 --token-embedding-type bf16 phi-4.bf16.gguf phi-4.bf16.q8.gguf q8_0
44
  llama-quantize --allow-requantize --pure phi-4.bf16.gguf phi-4.bf16.q8p.gguf q8_0
45
  ```
46
+
47
  ---------------------------------------------
48
 
49
  # Phi-4 Model Card