tzervas commited on
Commit
ef2c6a8
·
verified ·
1 Parent(s): 540ce91

Update model card with parameter count clarification

Browse files
Files changed (1) hide show
  1. README.md +17 -4
README.md CHANGED
@@ -8,24 +8,37 @@ tags:
8
  - 1.58-bit
9
  - phi-4
10
  - experimental
 
11
  library_name: safetensors
12
  pipeline_tag: text-generation
13
  language:
14
  - en
 
 
 
15
  ---
16
 
17
  # Phi-4-BitNet-1.58b
18
 
19
- BitNet 1.58-bit ternary quantization of Microsoft's Phi-4 14B model.
 
 
 
 
 
 
 
 
 
 
 
20
 
21
  ## Overview
22
 
23
- This is an **experimental** BitNet 1.58-bit quantization of the Phi-4 model using absmean scaling with group-wise quantization. The model stores weights as ternary values ({-1, 0, +1}) packed 4 values per byte.
24
 
25
  **This is research/experimental work. Quality and performance have not been formally benchmarked.**
26
 
27
- > **Note on Parameter Count**: HuggingFace may display a reduced parameter count because the quantized weights are packed (4 values per byte). The model retains the full 14B parameter architecture - only the weight storage is compressed.
28
-
29
  ## Specifications
30
 
31
  | Property | Value |
 
8
  - 1.58-bit
9
  - phi-4
10
  - experimental
11
+ - 14b-architecture
12
  library_name: safetensors
13
  pipeline_tag: text-generation
14
  language:
15
  - en
16
+ model_name: Phi-4-BitNet-1.58b
17
+ datasets: []
18
+ metrics: []
19
  ---
20
 
21
  # Phi-4-BitNet-1.58b
22
 
23
+ **Architecture: 14.7 Billion Parameters** | BitNet 1.58-bit Ternary Quantization
24
+
25
+ ---
26
+
27
+ > **IMPORTANT: Parameter Count Display**
28
+ >
29
+ > HuggingFace displays a reduced parameter count because it counts packed bytes, not actual parameters.
30
+ > This model has the **full 14.7B parameter Phi-4 architecture**.
31
+ > The weights are stored as ternary values ({-1, 0, +1}) packed 4 per byte, which reduces
32
+ > storage to 4.6 GB but preserves all 14.7 billion parameters.
33
+
34
+ ---
35
 
36
  ## Overview
37
 
38
+ This is an **experimental** BitNet 1.58-bit quantization of Microsoft's Phi-4 model using absmean scaling with group-wise quantization. The model stores weights as ternary values ({-1, 0, +1}) packed 4 values per byte.
39
 
40
  **This is research/experimental work. Quality and performance have not been formally benchmarked.**
41
 
 
 
42
  ## Specifications
43
 
44
  | Property | Value |