ASzecsenyi commited on
Commit
97fe4ee
·
verified ·
1 Parent(s): c3dc24e

Upload gpt_ts_d1/meta_000034.json with huggingface_hub

Browse files
Files changed (1) hide show
  1. gpt_ts_d1/meta_000034.json +96 -0
gpt_ts_d1/meta_000034.json ADDED
@@ -0,0 +1,96 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "step": 34,
3
+ "val_bpb": 3.227862918246376,
4
+ "model_config": {
5
+ "run": "gpt_ts_d1",
6
+ "device_type": "",
7
+ "depth": 1,
8
+ "aspect_ratio": 64,
9
+ "head_dim": 128,
10
+ "sequence_len": 256,
11
+ "vocab_size": 265,
12
+ "rva_blocks": [],
13
+ "gauss_blocks": [],
14
+ "num_iterations": -1,
15
+ "target_flops": -1.0,
16
+ "target_param_data_ratio": 200,
17
+ "data_dir": "tinystories_data",
18
+ "device_batch_size": 32,
19
+ "total_batch_size": 524288,
20
+ "embedding_lr": 0.3,
21
+ "unembedding_lr": 0.004,
22
+ "grad_clip": 1.0,
23
+ "weight_decay": 0.0,
24
+ "matrix_lr": 0.02,
25
+ "adam_beta1": 0.8,
26
+ "adam_beta2": 0.95,
27
+ "warmup_ratio": 0.0,
28
+ "warmdown_ratio": 0.4,
29
+ "final_lr_frac": 0.0,
30
+ "resume_from_step": -1,
31
+ "eval_every": 250,
32
+ "eval_tokens": 10485760,
33
+ "core_metric_every": -1,
34
+ "core_metric_max_per_task": 500,
35
+ "sample_every": -1,
36
+ "save_every": -1,
37
+ "push_checkpoints_to_hub": true,
38
+ "use_profiler": true,
39
+ "memory_history_max_entries": 100000,
40
+ "model_tag": "gpt_ts_d1",
41
+ "n_layer": 1,
42
+ "n_head": 1,
43
+ "n_kv_head": 1,
44
+ "n_embd": 64,
45
+ "d_state": 32
46
+ },
47
+ "user_config": {
48
+ "run": "gpt_ts_d1",
49
+ "device_type": "",
50
+ "depth": 1,
51
+ "aspect_ratio": 64,
52
+ "head_dim": 128,
53
+ "sequence_len": 256,
54
+ "vocab_size": 265,
55
+ "rva_blocks": [],
56
+ "gauss_blocks": [],
57
+ "num_iterations": -1,
58
+ "target_flops": -1.0,
59
+ "target_param_data_ratio": 200,
60
+ "data_dir": "tinystories_data",
61
+ "device_batch_size": 32,
62
+ "total_batch_size": 524288,
63
+ "embedding_lr": 0.3,
64
+ "unembedding_lr": 0.004,
65
+ "grad_clip": 1.0,
66
+ "weight_decay": 0.0,
67
+ "matrix_lr": 0.02,
68
+ "adam_beta1": 0.8,
69
+ "adam_beta2": 0.95,
70
+ "warmup_ratio": 0.0,
71
+ "warmdown_ratio": 0.4,
72
+ "final_lr_frac": 0.0,
73
+ "resume_from_step": -1,
74
+ "eval_every": 250,
75
+ "eval_tokens": 10485760,
76
+ "core_metric_every": -1,
77
+ "core_metric_max_per_task": 500,
78
+ "sample_every": -1,
79
+ "save_every": -1,
80
+ "push_checkpoints_to_hub": true,
81
+ "use_profiler": true,
82
+ "memory_history_max_entries": 100000,
83
+ "model_tag": "gpt_ts_d1"
84
+ },
85
+ "device_batch_size": 32,
86
+ "sequence_len": 256,
87
+ "dataloader_state_dict": {
88
+ "pq_idx": 0,
89
+ "rg_idx": 19
90
+ },
91
+ "loop_state": {
92
+ "min_val_bpb": 3.227862918246376,
93
+ "smooth_train_loss": 2.2646739822903337,
94
+ "total_training_time": 17.329952716827393
95
+ }
96
+ }