ASzecsenyi commited on
Commit
22dc06b
·
verified ·
1 Parent(s): d6b1909

Upload stage_1_ausvsb3b/meta_000813.json with huggingface_hub

Browse files
Files changed (1) hide show
  1. stage_1_ausvsb3b/meta_000813.json +142 -0
stage_1_ausvsb3b/meta_000813.json ADDED
@@ -0,0 +1,142 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "step": 813,
3
+ "val_bpb": 4.230076598531522,
4
+ "model_config": {
5
+ "run": "rva_2x1024/stage_1",
6
+ "wandb_group": "rva_2x1024",
7
+ "device_type": "cuda",
8
+ "config": "rva20m",
9
+ "depth": 2,
10
+ "aspect_ratio": 64,
11
+ "model_dim": 1024,
12
+ "head_dim": 64,
13
+ "sequence_len": 1024,
14
+ "vocab_size": 265,
15
+ "rva_blocks": [
16
+ 0,
17
+ 1
18
+ ],
19
+ "recurrent_vocab_sizes": [
20
+ 265,
21
+ 256
22
+ ],
23
+ "kla_blocks": [],
24
+ "mamba_blocks": [],
25
+ "gdn_blocks": [],
26
+ "gla_blocks": [],
27
+ "d_state": 16,
28
+ "mamba_params": true,
29
+ "kla_kernel": true,
30
+ "mimo_rank": 1,
31
+ "skip_around_kla": true,
32
+ "decoder_mlp": false,
33
+ "use_reparametrisation_trick": true,
34
+ "num_iterations": -1,
35
+ "target_flops": -1.0,
36
+ "target_param_data_ratio": 20,
37
+ "data_dir": "base_data",
38
+ "device_batch_size": 64,
39
+ "total_batch_size": 524288,
40
+ "use_muon": false,
41
+ "embedding_lr": 0.3,
42
+ "unembedding_lr": 0.004,
43
+ "grad_clip": 1.0,
44
+ "weight_decay": 0.0,
45
+ "matrix_lr": 0.02,
46
+ "adam_beta1": 0.8,
47
+ "adam_beta2": 0.95,
48
+ "warmup_ratio": 0.05,
49
+ "warmdown_ratio": 0.4,
50
+ "final_lr_frac": 0.0,
51
+ "resume_from_step": -1,
52
+ "eval_every": 250,
53
+ "eval_tokens": 10485760,
54
+ "core_metric_every": -1,
55
+ "core_metric_max_per_task": 500,
56
+ "sample_every": 250,
57
+ "save_every": 1000,
58
+ "push_checkpoints_to_hub": true,
59
+ "use_profiler": false,
60
+ "profile_step": 2,
61
+ "profile_micro_step": 0,
62
+ "memory_history_max_entries": 10000,
63
+ "model_tag": "rva_2x1024/stage_1",
64
+ "n_layer": 2,
65
+ "n_head": 16,
66
+ "n_kv_head": 16,
67
+ "n_embd": 1024
68
+ },
69
+ "user_config": {
70
+ "run": "rva_2x1024/stage_1",
71
+ "wandb_group": "rva_2x1024",
72
+ "device_type": "cuda",
73
+ "config": "rva20m",
74
+ "depth": 2,
75
+ "aspect_ratio": 64,
76
+ "model_dim": 1024,
77
+ "head_dim": 64,
78
+ "sequence_len": 1024,
79
+ "vocab_size": 265,
80
+ "rva_blocks": [
81
+ 0,
82
+ 1
83
+ ],
84
+ "recurrent_vocab_sizes": [
85
+ 265,
86
+ 256
87
+ ],
88
+ "kla_blocks": [],
89
+ "mamba_blocks": [],
90
+ "gdn_blocks": [],
91
+ "gla_blocks": [],
92
+ "d_state": 16,
93
+ "mamba_params": true,
94
+ "kla_kernel": true,
95
+ "mimo_rank": 1,
96
+ "skip_around_kla": true,
97
+ "decoder_mlp": false,
98
+ "use_reparametrisation_trick": true,
99
+ "num_iterations": -1,
100
+ "target_flops": -1.0,
101
+ "target_param_data_ratio": 20,
102
+ "data_dir": "base_data",
103
+ "device_batch_size": 64,
104
+ "total_batch_size": 524288,
105
+ "use_muon": false,
106
+ "embedding_lr": 0.3,
107
+ "unembedding_lr": 0.004,
108
+ "grad_clip": 1.0,
109
+ "weight_decay": 0.0,
110
+ "matrix_lr": 0.02,
111
+ "adam_beta1": 0.8,
112
+ "adam_beta2": 0.95,
113
+ "warmup_ratio": 0.05,
114
+ "warmdown_ratio": 0.4,
115
+ "final_lr_frac": 0.0,
116
+ "resume_from_step": -1,
117
+ "eval_every": 250,
118
+ "eval_tokens": 10485760,
119
+ "core_metric_every": -1,
120
+ "core_metric_max_per_task": 500,
121
+ "sample_every": 250,
122
+ "save_every": 1000,
123
+ "push_checkpoints_to_hub": true,
124
+ "use_profiler": false,
125
+ "profile_step": 2,
126
+ "profile_micro_step": 0,
127
+ "memory_history_max_entries": 10000,
128
+ "model_tag": "rva_2x1024/stage_1"
129
+ },
130
+ "device_batch_size": 64,
131
+ "sequence_len": 1024,
132
+ "dataloader_state_dict": {
133
+ "pq_idx": 1,
134
+ "rg_idx": 36
135
+ },
136
+ "loop_state": {
137
+ "min_val_bpb": 4.230076598531522,
138
+ "smooth_train_loss": 3.2408613999635754,
139
+ "smooth_aux_loss": 0.20064593989218568,
140
+ "total_training_time": 174.22333216667175
141
+ }
142
+ }