ASzecsenyi commited on
Commit
dbab6fc
·
verified ·
1 Parent(s): 2856273

Upload gpt-d1-ogxtv8si/meta_005000.json with huggingface_hub

Browse files
Files changed (1) hide show
  1. gpt-d1-ogxtv8si/meta_005000.json +74 -0
gpt-d1-ogxtv8si/meta_005000.json ADDED
@@ -0,0 +1,74 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "step": 5000,
3
+ "val_bpb": 2.052440584542226,
4
+ "model_config": {
5
+ "sequence_len": 2048,
6
+ "vocab_size": 265,
7
+ "n_layer": 1,
8
+ "n_head": 16,
9
+ "n_kv_head": 16,
10
+ "n_embd": 1024,
11
+ "adaptive_decay": false,
12
+ "rope": true,
13
+ "repetition_count": 1,
14
+ "mamba_layers": [],
15
+ "recurrent_vocab_layers": [],
16
+ "attention_time_decay": false
17
+ },
18
+ "user_config": {
19
+ "run": "gpt-d1",
20
+ "device_type": "cuda",
21
+ "depth": 1,
22
+ "repetition_count": 1,
23
+ "attention_time_decay": false,
24
+ "adaptive_decay": false,
25
+ "rope": true,
26
+ "init_from_tag": "",
27
+ "resume_from_tag": "",
28
+ "resume_from_step": -1,
29
+ "resume_from_hub": true,
30
+ "resume_hub_repo_id": "ASzecsenyi/VQLM",
31
+ "resume_hub_subdir": "",
32
+ "resume_hub_repo_type": "model",
33
+ "num_iterations": -1,
34
+ "target_flops": -1.0,
35
+ "target_param_data_ratio": 20,
36
+ "num_epochs": -1,
37
+ "data": "base_data",
38
+ "max_seq_len": 2048,
39
+ "device_batch_size": 4,
40
+ "total_batch_size": -1,
41
+ "embedding_lr": 0.2,
42
+ "unembedding_lr": 0.004,
43
+ "weight_decay": 0.0,
44
+ "matrix_lr": 0.02,
45
+ "base_lr": 1.0,
46
+ "grad_clip": 1.0,
47
+ "warmup_ratio": 0.05,
48
+ "warmdown_ratio": 0.2,
49
+ "final_lr_frac": 0.0,
50
+ "eval_every": 2500,
51
+ "eval_tokens": 10485760,
52
+ "core_metric_every": 2000,
53
+ "core_metric_max_per_task": 500,
54
+ "sample_every": 2500,
55
+ "checkpoint_every": -1,
56
+ "max_checkpoints": 3,
57
+ "push_checkpoints_to_hub": true,
58
+ "hf_repo_id": "ASzecsenyi/VQLM",
59
+ "hf_repo_type": "model",
60
+ "model_tag": "gpt-d1",
61
+ "ddp": true,
62
+ "ddp_rank": 0,
63
+ "ddp_local_rank": 0,
64
+ "ddp_world_size": 1
65
+ },
66
+ "device_batch_size": 4,
67
+ "max_seq_len": 2048,
68
+ "num_iterations": 32045,
69
+ "warmdown_ratio": 0.2,
70
+ "max_checkpoints": 3,
71
+ "push_checkpoints_to_hub": true,
72
+ "hf_repo_id": "ASzecsenyi/VQLM",
73
+ "hf_repo_type": "model"
74
+ }