AntonV HF Staff commited on
Commit
8db450a
·
verified ·
1 Parent(s): b25261b

Fixup unused attribtues and change to proper model type

Browse files
Files changed (1) hide show
  1. config.json +1 -12
config.json CHANGED
@@ -6,16 +6,6 @@
6
  "attention_dropout": 0.0,
7
  "dtype": "bfloat16",
8
  "eos_token_id": 100257,
9
- "fla_hybrid_attention_indices": [
10
- 3,
11
- 7,
12
- 11,
13
- 15,
14
- 19,
15
- 23,
16
- 27,
17
- 31
18
- ],
19
  "hidden_act": "silu",
20
  "hidden_size": 3840,
21
  "initializer_range": 0.02,
@@ -62,7 +52,7 @@
62
  "linear_use_gate": true,
63
  "linear_value_head_dim": 192,
64
  "max_position_embeddings": 8192,
65
- "model_type": "olmo3_2_hybrid",
66
  "num_attention_heads": 30,
67
  "num_hidden_layers": 32,
68
  "num_key_value_heads": 30,
@@ -72,7 +62,6 @@
72
  "rope_theta": 500000,
73
  "rope_type": "default"
74
  },
75
- "sliding_window": 4096,
76
  "tie_word_embeddings": false,
77
  "transformers_version": "5.0.0.dev0",
78
  "use_cache": true,
 
6
  "attention_dropout": 0.0,
7
  "dtype": "bfloat16",
8
  "eos_token_id": 100257,
 
 
 
 
 
 
 
 
 
 
9
  "hidden_act": "silu",
10
  "hidden_size": 3840,
11
  "initializer_range": 0.02,
 
52
  "linear_use_gate": true,
53
  "linear_value_head_dim": 192,
54
  "max_position_embeddings": 8192,
55
+ "model_type": "olmo_hybrid",
56
  "num_attention_heads": 30,
57
  "num_hidden_layers": 32,
58
  "num_key_value_heads": 30,
 
62
  "rope_theta": 500000,
63
  "rope_type": "default"
64
  },
 
65
  "tie_word_embeddings": false,
66
  "transformers_version": "5.0.0.dev0",
67
  "use_cache": true,