ASzecsenyi commited on
Commit
7e2ed1e
·
verified ·
1 Parent(s): e1140fb

Upload gpt_6x256_4dqmfbde/meta_001234.json with huggingface_hub

Browse files
Files changed (1) hide show
  1. gpt_6x256_4dqmfbde/meta_001234.json +147 -0
gpt_6x256_4dqmfbde/meta_001234.json ADDED
@@ -0,0 +1,147 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "step": 1234,
3
+ "val_bpb": 1.429377604643218,
4
+ "model_config": {
5
+ "run": "gpt_6x256",
6
+ "wandb_group": null,
7
+ "seed": 42,
8
+ "device_type": "cuda",
9
+ "config": "gpt5m",
10
+ "depth": 6,
11
+ "aspect_ratio": 64,
12
+ "model_dim": 256,
13
+ "head_dim": 128,
14
+ "sequence_len": 1024,
15
+ "vocab_size": 265,
16
+ "clustering_init_num_samples": 0,
17
+ "rva_blocks": [],
18
+ "recurrent_vocab_sizes": [],
19
+ "kla_blocks": [],
20
+ "mamba_blocks": [],
21
+ "gdn_blocks": [],
22
+ "gla_blocks": [],
23
+ "moe_blocks": [],
24
+ "d_state": 16,
25
+ "mamba_params": true,
26
+ "kla_kernel": true,
27
+ "mimo_rank": 1,
28
+ "skip_around_kla": true,
29
+ "decoder_mlp": false,
30
+ "use_reparametrisation_trick": true,
31
+ "num_experts": 4,
32
+ "moe_top_k": 2,
33
+ "moe_capacity_factor": 1.0,
34
+ "moe_aux_loss_weight": 0.01,
35
+ "moe_embedding_experts": false,
36
+ "num_iterations": 1234,
37
+ "target_flops": -1.0,
38
+ "target_param_data_ratio": 20,
39
+ "data_dir": "base_data",
40
+ "device_batch_size": 64,
41
+ "total_batch_size": 524288,
42
+ "use_muon": true,
43
+ "embedding_lr": 0.3,
44
+ "unembedding_lr": 0.004,
45
+ "grad_clip": 1.0,
46
+ "weight_decay": 0.0,
47
+ "matrix_lr": 0.02,
48
+ "adam_beta1": 0.8,
49
+ "adam_beta2": 0.95,
50
+ "warmup_ratio": 0.05,
51
+ "warmdown_ratio": 0.4,
52
+ "final_lr_frac": 0.0,
53
+ "resume_from_step": -1,
54
+ "eval_every": 250,
55
+ "eval_tokens": 10485760,
56
+ "core_metric_every": -1,
57
+ "core_metric_max_per_task": 500,
58
+ "sample_every": 250,
59
+ "save_every": 1000,
60
+ "push_checkpoints_to_hub": true,
61
+ "use_profiler": false,
62
+ "profile_step": 2,
63
+ "profile_micro_step": 0,
64
+ "memory_history_max_entries": 10000,
65
+ "model_tag": "gpt_6x256",
66
+ "n_layer": 6,
67
+ "n_head": 2,
68
+ "n_kv_head": 2,
69
+ "n_embd": 256
70
+ },
71
+ "user_config": {
72
+ "run": "gpt_6x256",
73
+ "wandb_group": null,
74
+ "seed": 42,
75
+ "device_type": "cuda",
76
+ "config": "gpt5m",
77
+ "depth": 6,
78
+ "aspect_ratio": 64,
79
+ "model_dim": 256,
80
+ "head_dim": 128,
81
+ "sequence_len": 1024,
82
+ "vocab_size": 265,
83
+ "clustering_init_num_samples": 0,
84
+ "rva_blocks": [],
85
+ "recurrent_vocab_sizes": [],
86
+ "kla_blocks": [],
87
+ "mamba_blocks": [],
88
+ "gdn_blocks": [],
89
+ "gla_blocks": [],
90
+ "moe_blocks": [],
91
+ "d_state": 16,
92
+ "mamba_params": true,
93
+ "kla_kernel": true,
94
+ "mimo_rank": 1,
95
+ "skip_around_kla": true,
96
+ "decoder_mlp": false,
97
+ "use_reparametrisation_trick": true,
98
+ "num_experts": 4,
99
+ "moe_top_k": 2,
100
+ "moe_capacity_factor": 1.0,
101
+ "moe_aux_loss_weight": 0.01,
102
+ "moe_embedding_experts": false,
103
+ "num_iterations": 1234,
104
+ "target_flops": -1.0,
105
+ "target_param_data_ratio": 20,
106
+ "data_dir": "base_data",
107
+ "device_batch_size": 64,
108
+ "total_batch_size": 524288,
109
+ "use_muon": true,
110
+ "embedding_lr": 0.3,
111
+ "unembedding_lr": 0.004,
112
+ "grad_clip": 1.0,
113
+ "weight_decay": 0.0,
114
+ "matrix_lr": 0.02,
115
+ "adam_beta1": 0.8,
116
+ "adam_beta2": 0.95,
117
+ "warmup_ratio": 0.05,
118
+ "warmdown_ratio": 0.4,
119
+ "final_lr_frac": 0.0,
120
+ "resume_from_step": -1,
121
+ "eval_every": 250,
122
+ "eval_tokens": 10485760,
123
+ "core_metric_every": -1,
124
+ "core_metric_max_per_task": 500,
125
+ "sample_every": 250,
126
+ "save_every": 1000,
127
+ "push_checkpoints_to_hub": true,
128
+ "use_profiler": false,
129
+ "profile_step": 2,
130
+ "profile_micro_step": 0,
131
+ "memory_history_max_entries": 10000,
132
+ "model_tag": "gpt_6x256",
133
+ "pod_name": "s2027538-infk8s-job-7dh6q-5pxnd"
134
+ },
135
+ "device_batch_size": 64,
136
+ "sequence_len": 1024,
137
+ "dataloader_state_dict": {
138
+ "pq_idx": 2,
139
+ "rg_idx": 28
140
+ },
141
+ "loop_state": {
142
+ "min_val_bpb": 1.429377604643218,
143
+ "smooth_train_loss": 1.0212040706081638,
144
+ "smooth_aux_loss": 0.0,
145
+ "total_training_time": 409.5275592803955
146
+ }
147
+ }