anthonym21 commited on
Commit
018e80e
·
verified ·
1 Parent(s): d46c8c6

Add config.json from base model

Browse files
Files changed (1) hide show
  1. config.json +15 -18
config.json CHANGED
@@ -1,28 +1,25 @@
1
  {
2
- "architecture": "Eve-2-MoE",
3
- "architectures": [
4
- "DeepSeekMoE"
5
- ],
6
  "auto_map": {
7
  "AutoConfig": "configuration_eve.EveConfig",
8
- "AutoModelForCausalLM": "modeling_eve.DeepSeekMoE"
9
  },
10
- "block_size": 2048,
11
- "dtype": "float32",
12
- "expert_intermediate_size": 1408,
13
- "head_dim": 64,
14
- "model_type": "eve_moe",
15
  "n_embd": 512,
16
  "n_head": 8,
17
- "n_layer": 12,
 
18
  "num_experts": 8,
19
- "rope_theta": 10000.0,
20
- "router_aux_loss_coef": 0.01,
21
- "shared_expert_intermediate_size": 1408,
22
  "top_k": 2,
23
- "torch_dtype": "bfloat16",
24
- "transformers_version": "5.1.0",
25
- "use_cache": false,
 
26
  "use_checkpointing": false,
27
- "vocab_size": 50304
 
 
 
28
  }
 
1
  {
2
+ "model_type": "eve-moe",
3
+ "architectures": ["EveMoEForCausalLM"],
 
 
4
  "auto_map": {
5
  "AutoConfig": "configuration_eve.EveConfig",
6
+ "AutoModelForCausalLM": "modeling_eve.EveMoEForCausalLM"
7
  },
8
+ "vocab_size": 50304,
9
+ "n_layer": 12,
 
 
 
10
  "n_embd": 512,
11
  "n_head": 8,
12
+ "head_dim": 64,
13
+ "block_size": 2048,
14
  "num_experts": 8,
 
 
 
15
  "top_k": 2,
16
+ "expert_intermediate_size": 1408,
17
+ "shared_expert_intermediate_size": 1408,
18
+ "router_aux_loss_coef": 0.01,
19
+ "rope_theta": 10000.0,
20
  "use_checkpointing": false,
21
+ "tie_word_embeddings": true,
22
+ "eos_token_id": 50256,
23
+ "torch_dtype": "bfloat16",
24
+ "transformers_version": "4.46.0"
25
  }