AdamF92 commited on
Commit
0e79b91
·
verified ·
1 Parent(s): c6a6132

In progress training - batch: 8192

Browse files
Files changed (2) hide show
  1. config.json +44 -0
  2. model.safetensors +3 -0
config.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "att_groups": 4,
3
+ "att_heads": 16,
4
+ "att_query_groups": 8,
5
+ "cross_att_type": "sqa",
6
+ "dense_layer_dim": 1536,
7
+ "embed_dim": 512,
8
+ "ff_activation": "silu",
9
+ "ff_dim": 192,
10
+ "ff_dropout": 0.0,
11
+ "final_stateless_layers_config": [
12
+ "moe",
13
+ "moe"
14
+ ],
15
+ "head_norm_type": "rms_norm",
16
+ "moe_bias_mode": "global",
17
+ "moe_grouped_gemm": true,
18
+ "moe_shared_experts_bias_mode": "global",
19
+ "moe_top_k": 10,
20
+ "moe_use_cutlass_grouped_gemm": true,
21
+ "moe_use_weighted_shared_experts": false,
22
+ "num_experts": 384,
23
+ "num_layers": 21,
24
+ "num_shared_experts": 2,
25
+ "rope_base": 100000,
26
+ "router_amp": true,
27
+ "self_att_type": "sqa",
28
+ "seq_len": 8192,
29
+ "shared_expert_dim": 384,
30
+ "stateless_layers_config": [
31
+ "dense",
32
+ "moe"
33
+ ],
34
+ "stm_size": 4096,
35
+ "use_attention_output_bias": false,
36
+ "use_flash_attention": true,
37
+ "use_gated": true,
38
+ "use_gated_attention": true,
39
+ "use_gated_cross_attention": false,
40
+ "use_head_norm": true,
41
+ "use_moe": true,
42
+ "use_vectorized_moe": true,
43
+ "vocab_size": 65536
44
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e3ad87cfc7fcc49e01bb9b649ffd2081d8f79aa1e200dd59bd6a2c24fb702cc
3
+ size 5728244528