weiweiz1 commited on
Commit
7c68236
·
verified ·
1 Parent(s): a8eac4f

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +5 -6
config.json CHANGED
@@ -1,4 +1,5 @@
1
  {
 
2
  "architectures": [
3
  "DeepseekV3ForCausalLM"
4
  ],
@@ -32,7 +33,6 @@
32
  "num_hidden_layers": 61,
33
  "num_key_value_heads": 128,
34
  "num_nextn_predict_layers": 1,
35
- "pad_token_id": 128815,
36
  "pretraining_tp": 1,
37
  "q_lora_rank": 1536,
38
  "qk_nope_head_dim": 128,
@@ -45,7 +45,7 @@
45
  "block_structure": null,
46
  "dynamic": "local",
47
  "group_size": 32,
48
- "is_mx": true,
49
  "num_bits": 8,
50
  "observer": "minmax",
51
  "observer_kwargs": {},
@@ -62,8 +62,8 @@
62
  "block_structure": null,
63
  "dynamic": false,
64
  "group_size": 32,
 
65
  "num_bits": 8,
66
- "is_mx": true,
67
  "observer": "minmax",
68
  "observer_kwargs": {},
69
  "strategy": "tensor_group",
@@ -72,7 +72,7 @@
72
  }
73
  }
74
  },
75
- "format": "nvfp4-pack-quantized",
76
  "global_compression_ratio": null,
77
  "ignore": [
78
  "lm_head"
@@ -99,8 +99,7 @@
99
  "topk_group": 4,
100
  "topk_method": "noaux_tc",
101
  "torch_dtype": "bfloat16",
102
- "transformers_version": "4.53.3",
103
- "unsloth_fixed": true,
104
  "use_cache": true,
105
  "v_head_dim": 128,
106
  "vocab_size": 129280
 
1
  {
2
+ "_name_or_path": "/data5/yliu7/HF_HOME/DeepSeek-R1-bf16/DeepSeek-R1-bf16",
3
  "architectures": [
4
  "DeepseekV3ForCausalLM"
5
  ],
 
33
  "num_hidden_layers": 61,
34
  "num_key_value_heads": 128,
35
  "num_nextn_predict_layers": 1,
 
36
  "pretraining_tp": 1,
37
  "q_lora_rank": 1536,
38
  "qk_nope_head_dim": 128,
 
45
  "block_structure": null,
46
  "dynamic": "local",
47
  "group_size": 32,
48
+ "is_mx": true,
49
  "num_bits": 8,
50
  "observer": "minmax",
51
  "observer_kwargs": {},
 
62
  "block_structure": null,
63
  "dynamic": false,
64
  "group_size": 32,
65
+ "is_mx": true,
66
  "num_bits": 8,
 
67
  "observer": "minmax",
68
  "observer_kwargs": {},
69
  "strategy": "tensor_group",
 
72
  }
73
  }
74
  },
75
+ "format": "float-quantized",
76
  "global_compression_ratio": null,
77
  "ignore": [
78
  "lm_head"
 
99
  "topk_group": 4,
100
  "topk_method": "noaux_tc",
101
  "torch_dtype": "bfloat16",
102
+ "transformers_version": "4.47.0",
 
103
  "use_cache": true,
104
  "v_head_dim": 128,
105
  "vocab_size": 129280