hiyouga commited on
Commit
83bb93b
·
verified ·
1 Parent(s): b2fccec

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +2 -6
config.json CHANGED
@@ -5,17 +5,12 @@
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
  "bos_token_id": 151643,
8
- "dtype": "bfloat16",
9
  "eos_token_id": 151645,
10
  "head_dim": 4,
11
  "hidden_act": "silu",
12
  "hidden_size": 16,
13
  "initializer_range": 0.02,
14
  "intermediate_size": 64,
15
- "layer_types": [
16
- "full_attention",
17
- "full_attention"
18
- ],
19
  "max_position_embeddings": 32768,
20
  "max_window_layers": 36,
21
  "model_type": "qwen3",
@@ -27,7 +22,8 @@
27
  "rope_theta": 5000000,
28
  "sliding_window": null,
29
  "tie_word_embeddings": true,
30
- "transformers_version": "4.57.1",
 
31
  "use_cache": true,
32
  "use_sliding_window": false,
33
  "vocab_size": 151936
 
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
  "bos_token_id": 151643,
 
8
  "eos_token_id": 151645,
9
  "head_dim": 4,
10
  "hidden_act": "silu",
11
  "hidden_size": 16,
12
  "initializer_range": 0.02,
13
  "intermediate_size": 64,
 
 
 
 
14
  "max_position_embeddings": 32768,
15
  "max_window_layers": 36,
16
  "model_type": "qwen3",
 
22
  "rope_theta": 5000000,
23
  "sliding_window": null,
24
  "tie_word_embeddings": true,
25
+ "torch_dtype": "bfloat16",
26
+ "transformers_version": "4.51.0",
27
  "use_cache": true,
28
  "use_sliding_window": false,
29
  "vocab_size": 151936